00:00:00.001 Started by upstream project "autotest-spdk-master-vs-dpdk-v22.11" build number 2381 00:00:00.001 originally caused by: 00:00:00.002 Started by upstream project "nightly-trigger" build number 3646 00:00:00.002 originally caused by: 00:00:00.002 Started by timer 00:00:00.002 Started by timer 00:00:00.020 Checking out git https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool into /var/jenkins_home/workspace/raid-vg-autotest_script/33b20b30f0a51e6b52980845e0f6aa336787973ad45e341fbbf98d1b65b265d4 to read jbp/jenkins/jjb-config/jobs/autotest-downstream/autotest-vg.groovy 00:00:00.021 The recommended git tool is: git 00:00:00.021 using credential 00000000-0000-0000-0000-000000000002 00:00:00.024 > git rev-parse --resolve-git-dir /var/jenkins_home/workspace/raid-vg-autotest_script/33b20b30f0a51e6b52980845e0f6aa336787973ad45e341fbbf98d1b65b265d4/jbp/.git # timeout=10 00:00:00.039 Fetching changes from the remote Git repository 00:00:00.041 > git config remote.origin.url https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool # timeout=10 00:00:00.056 Using shallow fetch with depth 1 00:00:00.056 Fetching upstream changes from https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool 00:00:00.056 > git --version # timeout=10 00:00:00.071 > git --version # 'git version 2.39.2' 00:00:00.071 using GIT_ASKPASS to set credentials SPDKCI HTTPS Credentials 00:00:00.092 Setting http proxy: proxy-dmz.intel.com:911 00:00:00.092 > git fetch --tags --force --progress --depth=1 -- https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool refs/heads/master # timeout=5 00:00:04.430 > git rev-parse origin/FETCH_HEAD^{commit} # timeout=10 00:00:04.443 > git rev-parse FETCH_HEAD^{commit} # timeout=10 00:00:04.456 Checking out Revision db4637e8b949f278f369ec13f70585206ccd9507 (FETCH_HEAD) 00:00:04.456 > git config core.sparsecheckout # timeout=10 00:00:04.470 > git read-tree -mu HEAD # timeout=10 00:00:04.486 > git checkout -f db4637e8b949f278f369ec13f70585206ccd9507 # timeout=5 00:00:04.513 Commit message: "jenkins/jjb-config: Add missing SPDK_TEST_NVME_INTERRUPT flag" 00:00:04.513 > git rev-list --no-walk db4637e8b949f278f369ec13f70585206ccd9507 # timeout=10 00:00:04.615 [Pipeline] Start of Pipeline 00:00:04.629 [Pipeline] library 00:00:04.631 Loading library shm_lib@master 00:00:04.631 Library shm_lib@master is cached. Copying from home. 00:00:04.645 [Pipeline] node 00:00:04.657 Running on VM-host-WFP7 in /var/jenkins/workspace/raid-vg-autotest 00:00:04.658 [Pipeline] { 00:00:04.665 [Pipeline] catchError 00:00:04.666 [Pipeline] { 00:00:04.677 [Pipeline] wrap 00:00:04.683 [Pipeline] { 00:00:04.689 [Pipeline] stage 00:00:04.690 [Pipeline] { (Prologue) 00:00:04.702 [Pipeline] echo 00:00:04.703 Node: VM-host-WFP7 00:00:04.707 [Pipeline] cleanWs 00:00:04.716 [WS-CLEANUP] Deleting project workspace... 00:00:04.716 [WS-CLEANUP] Deferred wipeout is used... 00:00:04.722 [WS-CLEANUP] done 00:00:04.884 [Pipeline] setCustomBuildProperty 00:00:04.971 [Pipeline] httpRequest 00:00:05.287 [Pipeline] echo 00:00:05.288 Sorcerer 10.211.164.20 is alive 00:00:05.295 [Pipeline] retry 00:00:05.296 [Pipeline] { 00:00:05.307 [Pipeline] httpRequest 00:00:05.311 HttpMethod: GET 00:00:05.312 URL: http://10.211.164.20/packages/jbp_db4637e8b949f278f369ec13f70585206ccd9507.tar.gz 00:00:05.312 Sending request to url: http://10.211.164.20/packages/jbp_db4637e8b949f278f369ec13f70585206ccd9507.tar.gz 00:00:05.316 Response Code: HTTP/1.1 200 OK 00:00:05.317 Success: Status code 200 is in the accepted range: 200,404 00:00:05.317 Saving response body to /var/jenkins/workspace/raid-vg-autotest/jbp_db4637e8b949f278f369ec13f70585206ccd9507.tar.gz 00:00:05.607 [Pipeline] } 00:00:05.623 [Pipeline] // retry 00:00:05.630 [Pipeline] sh 00:00:05.918 + tar --no-same-owner -xf jbp_db4637e8b949f278f369ec13f70585206ccd9507.tar.gz 00:00:05.936 [Pipeline] httpRequest 00:00:06.257 [Pipeline] echo 00:00:06.259 Sorcerer 10.211.164.20 is alive 00:00:06.269 [Pipeline] retry 00:00:06.272 [Pipeline] { 00:00:06.286 [Pipeline] httpRequest 00:00:06.291 HttpMethod: GET 00:00:06.292 URL: http://10.211.164.20/packages/spdk_dcc2ca8f30ea717d7f66cc9c92d44faa802d2c19.tar.gz 00:00:06.292 Sending request to url: http://10.211.164.20/packages/spdk_dcc2ca8f30ea717d7f66cc9c92d44faa802d2c19.tar.gz 00:00:06.296 Response Code: HTTP/1.1 200 OK 00:00:06.297 Success: Status code 200 is in the accepted range: 200,404 00:00:06.297 Saving response body to /var/jenkins/workspace/raid-vg-autotest/spdk_dcc2ca8f30ea717d7f66cc9c92d44faa802d2c19.tar.gz 00:01:56.629 [Pipeline] } 00:01:56.652 [Pipeline] // retry 00:01:56.661 [Pipeline] sh 00:01:56.968 + tar --no-same-owner -xf spdk_dcc2ca8f30ea717d7f66cc9c92d44faa802d2c19.tar.gz 00:02:00.268 [Pipeline] sh 00:02:00.553 + git -C spdk log --oneline -n5 00:02:00.553 dcc2ca8f3 bdev: fix per_channel data null when bdev_get_iostat with reset option 00:02:00.553 73f18e890 lib/reduce: fix the magic number of empty mapping detection. 00:02:00.553 029355612 bdev_ut: add manual examine bdev unit test case 00:02:00.553 fc96810c2 bdev: remove bdev from examine allow list on unregister 00:02:00.553 a0c128549 bdev/nvme: Make bdev nvme get and set opts APIs public 00:02:00.573 [Pipeline] withCredentials 00:02:00.587 > git --version # timeout=10 00:02:00.601 > git --version # 'git version 2.39.2' 00:02:00.620 Masking supported pattern matches of $GIT_PASSWORD or $GIT_ASKPASS 00:02:00.622 [Pipeline] { 00:02:00.631 [Pipeline] retry 00:02:00.633 [Pipeline] { 00:02:00.648 [Pipeline] sh 00:02:00.933 + git ls-remote http://dpdk.org/git/dpdk-stable v22.11.4 00:02:01.206 [Pipeline] } 00:02:01.224 [Pipeline] // retry 00:02:01.230 [Pipeline] } 00:02:01.246 [Pipeline] // withCredentials 00:02:01.256 [Pipeline] httpRequest 00:02:01.628 [Pipeline] echo 00:02:01.630 Sorcerer 10.211.164.20 is alive 00:02:01.641 [Pipeline] retry 00:02:01.643 [Pipeline] { 00:02:01.657 [Pipeline] httpRequest 00:02:01.662 HttpMethod: GET 00:02:01.663 URL: http://10.211.164.20/packages/dpdk_fee0f13c213d0584f0c42a51d0e0625d99a0b2f1.tar.gz 00:02:01.663 Sending request to url: http://10.211.164.20/packages/dpdk_fee0f13c213d0584f0c42a51d0e0625d99a0b2f1.tar.gz 00:02:01.665 Response Code: HTTP/1.1 200 OK 00:02:01.665 Success: Status code 200 is in the accepted range: 200,404 00:02:01.666 Saving response body to /var/jenkins/workspace/raid-vg-autotest/dpdk_fee0f13c213d0584f0c42a51d0e0625d99a0b2f1.tar.gz 00:02:08.282 [Pipeline] } 00:02:08.300 [Pipeline] // retry 00:02:08.309 [Pipeline] sh 00:02:08.594 + tar --no-same-owner -xf dpdk_fee0f13c213d0584f0c42a51d0e0625d99a0b2f1.tar.gz 00:02:09.988 [Pipeline] sh 00:02:10.272 + git -C dpdk log --oneline -n5 00:02:10.272 caf0f5d395 version: 22.11.4 00:02:10.272 7d6f1cc05f Revert "net/iavf: fix abnormal disable HW interrupt" 00:02:10.272 dc9c799c7d vhost: fix missing spinlock unlock 00:02:10.272 4307659a90 net/mlx5: fix LACP redirection in Rx domain 00:02:10.272 6ef77f2a5e net/gve: fix RX buffer size alignment 00:02:10.290 [Pipeline] writeFile 00:02:10.305 [Pipeline] sh 00:02:10.590 + jbp/jenkins/jjb-config/jobs/scripts/autorun_quirks.sh 00:02:10.602 [Pipeline] sh 00:02:10.885 + cat autorun-spdk.conf 00:02:10.885 SPDK_RUN_FUNCTIONAL_TEST=1 00:02:10.885 SPDK_RUN_ASAN=1 00:02:10.885 SPDK_RUN_UBSAN=1 00:02:10.885 SPDK_TEST_RAID=1 00:02:10.885 SPDK_TEST_NATIVE_DPDK=v22.11.4 00:02:10.885 SPDK_RUN_EXTERNAL_DPDK=/home/vagrant/spdk_repo/dpdk/build 00:02:10.885 SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:02:10.892 RUN_NIGHTLY=1 00:02:10.894 [Pipeline] } 00:02:10.910 [Pipeline] // stage 00:02:10.925 [Pipeline] stage 00:02:10.928 [Pipeline] { (Run VM) 00:02:10.942 [Pipeline] sh 00:02:11.224 + jbp/jenkins/jjb-config/jobs/scripts/prepare_nvme.sh 00:02:11.224 + echo 'Start stage prepare_nvme.sh' 00:02:11.224 Start stage prepare_nvme.sh 00:02:11.224 + [[ -n 1 ]] 00:02:11.224 + disk_prefix=ex1 00:02:11.224 + [[ -n /var/jenkins/workspace/raid-vg-autotest ]] 00:02:11.224 + [[ -e /var/jenkins/workspace/raid-vg-autotest/autorun-spdk.conf ]] 00:02:11.224 + source /var/jenkins/workspace/raid-vg-autotest/autorun-spdk.conf 00:02:11.224 ++ SPDK_RUN_FUNCTIONAL_TEST=1 00:02:11.224 ++ SPDK_RUN_ASAN=1 00:02:11.224 ++ SPDK_RUN_UBSAN=1 00:02:11.224 ++ SPDK_TEST_RAID=1 00:02:11.224 ++ SPDK_TEST_NATIVE_DPDK=v22.11.4 00:02:11.224 ++ SPDK_RUN_EXTERNAL_DPDK=/home/vagrant/spdk_repo/dpdk/build 00:02:11.224 ++ SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:02:11.224 ++ RUN_NIGHTLY=1 00:02:11.224 + cd /var/jenkins/workspace/raid-vg-autotest 00:02:11.224 + nvme_files=() 00:02:11.224 + declare -A nvme_files 00:02:11.224 + backend_dir=/var/lib/libvirt/images/backends 00:02:11.224 + nvme_files['nvme.img']=5G 00:02:11.224 + nvme_files['nvme-cmb.img']=5G 00:02:11.224 + nvme_files['nvme-multi0.img']=4G 00:02:11.224 + nvme_files['nvme-multi1.img']=4G 00:02:11.224 + nvme_files['nvme-multi2.img']=4G 00:02:11.224 + nvme_files['nvme-openstack.img']=8G 00:02:11.224 + nvme_files['nvme-zns.img']=5G 00:02:11.224 + (( SPDK_TEST_NVME_PMR == 1 )) 00:02:11.224 + (( SPDK_TEST_FTL == 1 )) 00:02:11.224 + (( SPDK_TEST_NVME_FDP == 1 )) 00:02:11.224 + [[ ! -d /var/lib/libvirt/images/backends ]] 00:02:11.224 + for nvme in "${!nvme_files[@]}" 00:02:11.224 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex1-nvme-multi2.img -s 4G 00:02:11.224 Formatting '/var/lib/libvirt/images/backends/ex1-nvme-multi2.img', fmt=raw size=4294967296 preallocation=falloc 00:02:11.224 + for nvme in "${!nvme_files[@]}" 00:02:11.224 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex1-nvme-cmb.img -s 5G 00:02:11.224 Formatting '/var/lib/libvirt/images/backends/ex1-nvme-cmb.img', fmt=raw size=5368709120 preallocation=falloc 00:02:11.224 + for nvme in "${!nvme_files[@]}" 00:02:11.224 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex1-nvme-openstack.img -s 8G 00:02:11.224 Formatting '/var/lib/libvirt/images/backends/ex1-nvme-openstack.img', fmt=raw size=8589934592 preallocation=falloc 00:02:11.224 + for nvme in "${!nvme_files[@]}" 00:02:11.224 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex1-nvme-zns.img -s 5G 00:02:11.224 Formatting '/var/lib/libvirt/images/backends/ex1-nvme-zns.img', fmt=raw size=5368709120 preallocation=falloc 00:02:11.224 + for nvme in "${!nvme_files[@]}" 00:02:11.224 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex1-nvme-multi1.img -s 4G 00:02:11.224 Formatting '/var/lib/libvirt/images/backends/ex1-nvme-multi1.img', fmt=raw size=4294967296 preallocation=falloc 00:02:11.224 + for nvme in "${!nvme_files[@]}" 00:02:11.224 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex1-nvme-multi0.img -s 4G 00:02:11.224 Formatting '/var/lib/libvirt/images/backends/ex1-nvme-multi0.img', fmt=raw size=4294967296 preallocation=falloc 00:02:11.224 + for nvme in "${!nvme_files[@]}" 00:02:11.224 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex1-nvme.img -s 5G 00:02:11.484 Formatting '/var/lib/libvirt/images/backends/ex1-nvme.img', fmt=raw size=5368709120 preallocation=falloc 00:02:11.484 ++ sudo grep -rl ex1-nvme.img /etc/libvirt/qemu 00:02:11.484 + echo 'End stage prepare_nvme.sh' 00:02:11.484 End stage prepare_nvme.sh 00:02:11.496 [Pipeline] sh 00:02:11.783 + DISTRO=fedora39 CPUS=10 RAM=12288 jbp/jenkins/jjb-config/jobs/scripts/vagrant_create_vm.sh 00:02:11.783 Setup: -n 10 -s 12288 -x http://proxy-dmz.intel.com:911 -p libvirt --qemu-emulator=/usr/local/qemu/vanilla-v8.0.0/bin/qemu-system-x86_64 -b /var/lib/libvirt/images/backends/ex1-nvme.img -b /var/lib/libvirt/images/backends/ex1-nvme-multi0.img,nvme,/var/lib/libvirt/images/backends/ex1-nvme-multi1.img:/var/lib/libvirt/images/backends/ex1-nvme-multi2.img -H -a -v -f fedora39 00:02:11.783 00:02:11.783 DIR=/var/jenkins/workspace/raid-vg-autotest/spdk/scripts/vagrant 00:02:11.783 SPDK_DIR=/var/jenkins/workspace/raid-vg-autotest/spdk 00:02:11.783 VAGRANT_TARGET=/var/jenkins/workspace/raid-vg-autotest 00:02:11.783 HELP=0 00:02:11.783 DRY_RUN=0 00:02:11.783 NVME_FILE=/var/lib/libvirt/images/backends/ex1-nvme.img,/var/lib/libvirt/images/backends/ex1-nvme-multi0.img, 00:02:11.783 NVME_DISKS_TYPE=nvme,nvme, 00:02:11.783 NVME_AUTO_CREATE=0 00:02:11.783 NVME_DISKS_NAMESPACES=,/var/lib/libvirt/images/backends/ex1-nvme-multi1.img:/var/lib/libvirt/images/backends/ex1-nvme-multi2.img, 00:02:11.783 NVME_CMB=,, 00:02:11.783 NVME_PMR=,, 00:02:11.783 NVME_ZNS=,, 00:02:11.783 NVME_MS=,, 00:02:11.783 NVME_FDP=,, 00:02:11.783 SPDK_VAGRANT_DISTRO=fedora39 00:02:11.783 SPDK_VAGRANT_VMCPU=10 00:02:11.783 SPDK_VAGRANT_VMRAM=12288 00:02:11.783 SPDK_VAGRANT_PROVIDER=libvirt 00:02:11.783 SPDK_VAGRANT_HTTP_PROXY=http://proxy-dmz.intel.com:911 00:02:11.783 SPDK_QEMU_EMULATOR=/usr/local/qemu/vanilla-v8.0.0/bin/qemu-system-x86_64 00:02:11.783 SPDK_OPENSTACK_NETWORK=0 00:02:11.783 VAGRANT_PACKAGE_BOX=0 00:02:11.783 VAGRANTFILE=/var/jenkins/workspace/raid-vg-autotest/spdk/scripts/vagrant/Vagrantfile 00:02:11.783 FORCE_DISTRO=true 00:02:11.783 VAGRANT_BOX_VERSION= 00:02:11.783 EXTRA_VAGRANTFILES= 00:02:11.783 NIC_MODEL=virtio 00:02:11.783 00:02:11.783 mkdir: created directory '/var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt' 00:02:11.783 /var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt /var/jenkins/workspace/raid-vg-autotest 00:02:13.692 Bringing machine 'default' up with 'libvirt' provider... 00:02:14.262 ==> default: Creating image (snapshot of base box volume). 00:02:14.262 ==> default: Creating domain with the following settings... 00:02:14.262 ==> default: -- Name: fedora39-39-1.5-1721788873-2326_default_1732028884_f7c4a23ed3fa72c7d0b8 00:02:14.262 ==> default: -- Domain type: kvm 00:02:14.262 ==> default: -- Cpus: 10 00:02:14.262 ==> default: -- Feature: acpi 00:02:14.262 ==> default: -- Feature: apic 00:02:14.262 ==> default: -- Feature: pae 00:02:14.262 ==> default: -- Memory: 12288M 00:02:14.262 ==> default: -- Memory Backing: hugepages: 00:02:14.262 ==> default: -- Management MAC: 00:02:14.262 ==> default: -- Loader: 00:02:14.262 ==> default: -- Nvram: 00:02:14.262 ==> default: -- Base box: spdk/fedora39 00:02:14.262 ==> default: -- Storage pool: default 00:02:14.262 ==> default: -- Image: /var/lib/libvirt/images/fedora39-39-1.5-1721788873-2326_default_1732028884_f7c4a23ed3fa72c7d0b8.img (20G) 00:02:14.262 ==> default: -- Volume Cache: default 00:02:14.262 ==> default: -- Kernel: 00:02:14.262 ==> default: -- Initrd: 00:02:14.262 ==> default: -- Graphics Type: vnc 00:02:14.262 ==> default: -- Graphics Port: -1 00:02:14.262 ==> default: -- Graphics IP: 127.0.0.1 00:02:14.262 ==> default: -- Graphics Password: Not defined 00:02:14.262 ==> default: -- Video Type: cirrus 00:02:14.262 ==> default: -- Video VRAM: 9216 00:02:14.262 ==> default: -- Sound Type: 00:02:14.262 ==> default: -- Keymap: en-us 00:02:14.262 ==> default: -- TPM Path: 00:02:14.262 ==> default: -- INPUT: type=mouse, bus=ps2 00:02:14.262 ==> default: -- Command line args: 00:02:14.262 ==> default: -> value=-device, 00:02:14.262 ==> default: -> value=nvme,id=nvme-0,serial=12340,addr=0x10, 00:02:14.262 ==> default: -> value=-drive, 00:02:14.262 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex1-nvme.img,if=none,id=nvme-0-drive0, 00:02:14.262 ==> default: -> value=-device, 00:02:14.262 ==> default: -> value=nvme-ns,drive=nvme-0-drive0,bus=nvme-0,nsid=1,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:02:14.262 ==> default: -> value=-device, 00:02:14.262 ==> default: -> value=nvme,id=nvme-1,serial=12341,addr=0x11, 00:02:14.262 ==> default: -> value=-drive, 00:02:14.262 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex1-nvme-multi0.img,if=none,id=nvme-1-drive0, 00:02:14.262 ==> default: -> value=-device, 00:02:14.262 ==> default: -> value=nvme-ns,drive=nvme-1-drive0,bus=nvme-1,nsid=1,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:02:14.262 ==> default: -> value=-drive, 00:02:14.262 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex1-nvme-multi1.img,if=none,id=nvme-1-drive1, 00:02:14.262 ==> default: -> value=-device, 00:02:14.262 ==> default: -> value=nvme-ns,drive=nvme-1-drive1,bus=nvme-1,nsid=2,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:02:14.262 ==> default: -> value=-drive, 00:02:14.262 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex1-nvme-multi2.img,if=none,id=nvme-1-drive2, 00:02:14.262 ==> default: -> value=-device, 00:02:14.262 ==> default: -> value=nvme-ns,drive=nvme-1-drive2,bus=nvme-1,nsid=3,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:02:14.522 ==> default: Creating shared folders metadata... 00:02:14.522 ==> default: Starting domain. 00:02:15.904 ==> default: Waiting for domain to get an IP address... 00:02:34.004 ==> default: Waiting for SSH to become available... 00:02:34.004 ==> default: Configuring and enabling network interfaces... 00:02:39.273 default: SSH address: 192.168.121.125:22 00:02:39.273 default: SSH username: vagrant 00:02:39.273 default: SSH auth method: private key 00:02:41.171 ==> default: Rsyncing folder: /mnt/jenkins_nvme/jenkins/workspace/raid-vg-autotest/spdk/ => /home/vagrant/spdk_repo/spdk 00:02:51.151 ==> default: Rsyncing folder: /mnt/jenkins_nvme/jenkins/workspace/raid-vg-autotest/dpdk/ => /home/vagrant/spdk_repo/dpdk 00:02:56.421 ==> default: Mounting SSHFS shared folder... 00:02:58.960 ==> default: Mounting folder via SSHFS: /mnt/jenkins_nvme/jenkins/workspace/raid-vg-autotest/fedora39-libvirt/output => /home/vagrant/spdk_repo/output 00:02:58.960 ==> default: Checking Mount.. 00:03:00.339 ==> default: Folder Successfully Mounted! 00:03:00.339 ==> default: Running provisioner: file... 00:03:01.787 default: ~/.gitconfig => .gitconfig 00:03:02.046 00:03:02.046 SUCCESS! 00:03:02.046 00:03:02.046 cd to /var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt and type "vagrant ssh" to use. 00:03:02.046 Use vagrant "suspend" and vagrant "resume" to stop and start. 00:03:02.046 Use vagrant "destroy" followed by "rm -rf /var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt" to destroy all trace of vm. 00:03:02.046 00:03:02.055 [Pipeline] } 00:03:02.070 [Pipeline] // stage 00:03:02.079 [Pipeline] dir 00:03:02.079 Running in /var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt 00:03:02.081 [Pipeline] { 00:03:02.093 [Pipeline] catchError 00:03:02.095 [Pipeline] { 00:03:02.107 [Pipeline] sh 00:03:02.388 + vagrant ssh-config --host+ vagrant 00:03:02.388 sed -ne /^Host/,$p 00:03:02.388 + tee ssh_conf 00:03:04.922 Host vagrant 00:03:04.922 HostName 192.168.121.125 00:03:04.922 User vagrant 00:03:04.922 Port 22 00:03:04.922 UserKnownHostsFile /dev/null 00:03:04.922 StrictHostKeyChecking no 00:03:04.922 PasswordAuthentication no 00:03:04.922 IdentityFile /var/lib/libvirt/images/.vagrant.d/boxes/spdk-VAGRANTSLASH-fedora39/39-1.5-1721788873-2326/libvirt/fedora39 00:03:04.922 IdentitiesOnly yes 00:03:04.922 LogLevel FATAL 00:03:04.922 ForwardAgent yes 00:03:04.922 ForwardX11 yes 00:03:04.922 00:03:04.936 [Pipeline] withEnv 00:03:04.938 [Pipeline] { 00:03:04.953 [Pipeline] sh 00:03:05.235 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant #!/bin/bash 00:03:05.235 source /etc/os-release 00:03:05.235 [[ -e /image.version ]] && img=$(< /image.version) 00:03:05.235 # Minimal, systemd-like check. 00:03:05.235 if [[ -e /.dockerenv ]]; then 00:03:05.235 # Clear garbage from the node's name: 00:03:05.235 # agt-er_autotest_547-896 -> autotest_547-896 00:03:05.235 # $HOSTNAME is the actual container id 00:03:05.235 agent=$HOSTNAME@${DOCKER_SWARM_PLUGIN_JENKINS_AGENT_NAME#*_} 00:03:05.235 if grep -q "/etc/hostname" /proc/self/mountinfo; then 00:03:05.235 # We can assume this is a mount from a host where container is running, 00:03:05.235 # so fetch its hostname to easily identify the target swarm worker. 00:03:05.235 container="$(< /etc/hostname) ($agent)" 00:03:05.235 else 00:03:05.235 # Fallback 00:03:05.235 container=$agent 00:03:05.235 fi 00:03:05.235 fi 00:03:05.235 echo "${NAME} ${VERSION_ID}|$(uname -r)|${img:-N/A}|${container:-N/A}" 00:03:05.235 00:03:05.504 [Pipeline] } 00:03:05.520 [Pipeline] // withEnv 00:03:05.530 [Pipeline] setCustomBuildProperty 00:03:05.544 [Pipeline] stage 00:03:05.546 [Pipeline] { (Tests) 00:03:05.563 [Pipeline] sh 00:03:05.846 + scp -F ssh_conf -r /var/jenkins/workspace/raid-vg-autotest/jbp/jenkins/jjb-config/jobs/scripts/autoruner.sh vagrant@vagrant:./ 00:03:06.137 [Pipeline] sh 00:03:06.418 + scp -F ssh_conf -r /var/jenkins/workspace/raid-vg-autotest/jbp/jenkins/jjb-config/jobs/scripts/pkgdep-autoruner.sh vagrant@vagrant:./ 00:03:06.691 [Pipeline] timeout 00:03:06.691 Timeout set to expire in 1 hr 30 min 00:03:06.693 [Pipeline] { 00:03:06.706 [Pipeline] sh 00:03:06.987 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant git -C spdk_repo/spdk reset --hard 00:03:07.555 HEAD is now at dcc2ca8f3 bdev: fix per_channel data null when bdev_get_iostat with reset option 00:03:07.568 [Pipeline] sh 00:03:07.850 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant sudo chown vagrant:vagrant spdk_repo 00:03:08.122 [Pipeline] sh 00:03:08.403 + scp -F ssh_conf -r /var/jenkins/workspace/raid-vg-autotest/autorun-spdk.conf vagrant@vagrant:spdk_repo 00:03:08.676 [Pipeline] sh 00:03:08.954 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant JOB_BASE_NAME=raid-vg-autotest ./autoruner.sh spdk_repo 00:03:09.212 ++ readlink -f spdk_repo 00:03:09.212 + DIR_ROOT=/home/vagrant/spdk_repo 00:03:09.212 + [[ -n /home/vagrant/spdk_repo ]] 00:03:09.212 + DIR_SPDK=/home/vagrant/spdk_repo/spdk 00:03:09.212 + DIR_OUTPUT=/home/vagrant/spdk_repo/output 00:03:09.212 + [[ -d /home/vagrant/spdk_repo/spdk ]] 00:03:09.212 + [[ ! -d /home/vagrant/spdk_repo/output ]] 00:03:09.212 + [[ -d /home/vagrant/spdk_repo/output ]] 00:03:09.212 + [[ raid-vg-autotest == pkgdep-* ]] 00:03:09.212 + cd /home/vagrant/spdk_repo 00:03:09.212 + source /etc/os-release 00:03:09.212 ++ NAME='Fedora Linux' 00:03:09.212 ++ VERSION='39 (Cloud Edition)' 00:03:09.212 ++ ID=fedora 00:03:09.212 ++ VERSION_ID=39 00:03:09.212 ++ VERSION_CODENAME= 00:03:09.212 ++ PLATFORM_ID=platform:f39 00:03:09.212 ++ PRETTY_NAME='Fedora Linux 39 (Cloud Edition)' 00:03:09.212 ++ ANSI_COLOR='0;38;2;60;110;180' 00:03:09.212 ++ LOGO=fedora-logo-icon 00:03:09.212 ++ CPE_NAME=cpe:/o:fedoraproject:fedora:39 00:03:09.212 ++ HOME_URL=https://fedoraproject.org/ 00:03:09.212 ++ DOCUMENTATION_URL=https://docs.fedoraproject.org/en-US/fedora/f39/system-administrators-guide/ 00:03:09.212 ++ SUPPORT_URL=https://ask.fedoraproject.org/ 00:03:09.212 ++ BUG_REPORT_URL=https://bugzilla.redhat.com/ 00:03:09.212 ++ REDHAT_BUGZILLA_PRODUCT=Fedora 00:03:09.212 ++ REDHAT_BUGZILLA_PRODUCT_VERSION=39 00:03:09.212 ++ REDHAT_SUPPORT_PRODUCT=Fedora 00:03:09.212 ++ REDHAT_SUPPORT_PRODUCT_VERSION=39 00:03:09.212 ++ SUPPORT_END=2024-11-12 00:03:09.212 ++ VARIANT='Cloud Edition' 00:03:09.212 ++ VARIANT_ID=cloud 00:03:09.212 + uname -a 00:03:09.212 Linux fedora39-cloud-1721788873-2326 6.8.9-200.fc39.x86_64 #1 SMP PREEMPT_DYNAMIC Wed Jul 24 03:04:40 UTC 2024 x86_64 GNU/Linux 00:03:09.212 + sudo /home/vagrant/spdk_repo/spdk/scripts/setup.sh status 00:03:09.778 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:03:09.778 Hugepages 00:03:09.778 node hugesize free / total 00:03:09.779 node0 1048576kB 0 / 0 00:03:09.779 node0 2048kB 0 / 0 00:03:09.779 00:03:09.779 Type BDF Vendor Device NUMA Driver Device Block devices 00:03:09.779 virtio 0000:00:03.0 1af4 1001 unknown virtio-pci - vda 00:03:09.779 NVMe 0000:00:10.0 1b36 0010 unknown nvme nvme0 nvme0n1 00:03:09.779 NVMe 0000:00:11.0 1b36 0010 unknown nvme nvme1 nvme1n1 nvme1n2 nvme1n3 00:03:09.779 + rm -f /tmp/spdk-ld-path 00:03:09.779 + source autorun-spdk.conf 00:03:09.779 ++ SPDK_RUN_FUNCTIONAL_TEST=1 00:03:09.779 ++ SPDK_RUN_ASAN=1 00:03:09.779 ++ SPDK_RUN_UBSAN=1 00:03:09.779 ++ SPDK_TEST_RAID=1 00:03:09.779 ++ SPDK_TEST_NATIVE_DPDK=v22.11.4 00:03:09.779 ++ SPDK_RUN_EXTERNAL_DPDK=/home/vagrant/spdk_repo/dpdk/build 00:03:09.779 ++ SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:03:09.779 ++ RUN_NIGHTLY=1 00:03:09.779 + (( SPDK_TEST_NVME_CMB == 1 || SPDK_TEST_NVME_PMR == 1 )) 00:03:09.779 + [[ -n '' ]] 00:03:09.779 + sudo git config --global --add safe.directory /home/vagrant/spdk_repo/spdk 00:03:09.779 + for M in /var/spdk/build-*-manifest.txt 00:03:09.779 + [[ -f /var/spdk/build-kernel-manifest.txt ]] 00:03:09.779 + cp /var/spdk/build-kernel-manifest.txt /home/vagrant/spdk_repo/output/ 00:03:09.779 + for M in /var/spdk/build-*-manifest.txt 00:03:09.779 + [[ -f /var/spdk/build-pkg-manifest.txt ]] 00:03:09.779 + cp /var/spdk/build-pkg-manifest.txt /home/vagrant/spdk_repo/output/ 00:03:09.779 + for M in /var/spdk/build-*-manifest.txt 00:03:09.779 + [[ -f /var/spdk/build-repo-manifest.txt ]] 00:03:09.779 + cp /var/spdk/build-repo-manifest.txt /home/vagrant/spdk_repo/output/ 00:03:09.779 ++ uname 00:03:09.779 + [[ Linux == \L\i\n\u\x ]] 00:03:09.779 + sudo dmesg -T 00:03:10.037 + sudo dmesg --clear 00:03:10.037 + dmesg_pid=6164 00:03:10.037 + sudo dmesg -Tw 00:03:10.037 + [[ Fedora Linux == FreeBSD ]] 00:03:10.037 + export UNBIND_ENTIRE_IOMMU_GROUP=yes 00:03:10.037 + UNBIND_ENTIRE_IOMMU_GROUP=yes 00:03:10.037 + [[ -e /var/spdk/dependencies/vhost/spdk_test_image.qcow2 ]] 00:03:10.037 + [[ -x /usr/src/fio-static/fio ]] 00:03:10.037 + export FIO_BIN=/usr/src/fio-static/fio 00:03:10.037 + FIO_BIN=/usr/src/fio-static/fio 00:03:10.037 + [[ '' == \/\q\e\m\u\_\v\f\i\o\/* ]] 00:03:10.037 + [[ ! -v VFIO_QEMU_BIN ]] 00:03:10.037 + [[ -e /usr/local/qemu/vfio-user-latest ]] 00:03:10.037 + export VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:03:10.037 + VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:03:10.037 + [[ -e /usr/local/qemu/vanilla-latest ]] 00:03:10.037 + export QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:03:10.037 + QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:03:10.037 + spdk/autorun.sh /home/vagrant/spdk_repo/autorun-spdk.conf 00:03:10.037 15:09:00 -- common/autotest_common.sh@1692 -- $ [[ n == y ]] 00:03:10.037 15:09:00 -- spdk/autorun.sh@20 -- $ source /home/vagrant/spdk_repo/autorun-spdk.conf 00:03:10.037 15:09:00 -- spdk_repo/autorun-spdk.conf@1 -- $ SPDK_RUN_FUNCTIONAL_TEST=1 00:03:10.037 15:09:00 -- spdk_repo/autorun-spdk.conf@2 -- $ SPDK_RUN_ASAN=1 00:03:10.037 15:09:00 -- spdk_repo/autorun-spdk.conf@3 -- $ SPDK_RUN_UBSAN=1 00:03:10.037 15:09:00 -- spdk_repo/autorun-spdk.conf@4 -- $ SPDK_TEST_RAID=1 00:03:10.037 15:09:00 -- spdk_repo/autorun-spdk.conf@5 -- $ SPDK_TEST_NATIVE_DPDK=v22.11.4 00:03:10.037 15:09:00 -- spdk_repo/autorun-spdk.conf@6 -- $ SPDK_RUN_EXTERNAL_DPDK=/home/vagrant/spdk_repo/dpdk/build 00:03:10.037 15:09:00 -- spdk_repo/autorun-spdk.conf@7 -- $ SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:03:10.037 15:09:00 -- spdk_repo/autorun-spdk.conf@8 -- $ RUN_NIGHTLY=1 00:03:10.037 15:09:00 -- spdk/autorun.sh@22 -- $ trap 'timing_finish || exit 1' EXIT 00:03:10.037 15:09:00 -- spdk/autorun.sh@25 -- $ /home/vagrant/spdk_repo/spdk/autobuild.sh /home/vagrant/spdk_repo/autorun-spdk.conf 00:03:10.038 15:09:00 -- common/autotest_common.sh@1692 -- $ [[ n == y ]] 00:03:10.038 15:09:00 -- common/autobuild_common.sh@15 -- $ source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:03:10.038 15:09:00 -- scripts/common.sh@15 -- $ shopt -s extglob 00:03:10.038 15:09:00 -- scripts/common.sh@544 -- $ [[ -e /bin/wpdk_common.sh ]] 00:03:10.038 15:09:00 -- scripts/common.sh@552 -- $ [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:03:10.038 15:09:00 -- scripts/common.sh@553 -- $ source /etc/opt/spdk-pkgdep/paths/export.sh 00:03:10.038 15:09:00 -- paths/export.sh@2 -- $ PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:03:10.038 15:09:00 -- paths/export.sh@3 -- $ PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:03:10.038 15:09:00 -- paths/export.sh@4 -- $ PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:03:10.038 15:09:00 -- paths/export.sh@5 -- $ export PATH 00:03:10.038 15:09:00 -- paths/export.sh@6 -- $ echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:03:10.038 15:09:00 -- common/autobuild_common.sh@485 -- $ out=/home/vagrant/spdk_repo/spdk/../output 00:03:10.038 15:09:00 -- common/autobuild_common.sh@486 -- $ date +%s 00:03:10.296 15:09:00 -- common/autobuild_common.sh@486 -- $ mktemp -dt spdk_1732028940.XXXXXX 00:03:10.296 15:09:00 -- common/autobuild_common.sh@486 -- $ SPDK_WORKSPACE=/tmp/spdk_1732028940.bNgZsl 00:03:10.296 15:09:00 -- common/autobuild_common.sh@488 -- $ [[ -n '' ]] 00:03:10.296 15:09:00 -- common/autobuild_common.sh@492 -- $ '[' -n v22.11.4 ']' 00:03:10.296 15:09:00 -- common/autobuild_common.sh@493 -- $ dirname /home/vagrant/spdk_repo/dpdk/build 00:03:10.296 15:09:00 -- common/autobuild_common.sh@493 -- $ scanbuild_exclude=' --exclude /home/vagrant/spdk_repo/dpdk' 00:03:10.296 15:09:00 -- common/autobuild_common.sh@499 -- $ scanbuild_exclude+=' --exclude /home/vagrant/spdk_repo/spdk/xnvme --exclude /tmp' 00:03:10.296 15:09:00 -- common/autobuild_common.sh@501 -- $ scanbuild='scan-build -o /home/vagrant/spdk_repo/spdk/../output/scan-build-tmp --exclude /home/vagrant/spdk_repo/dpdk --exclude /home/vagrant/spdk_repo/spdk/xnvme --exclude /tmp --status-bugs' 00:03:10.296 15:09:00 -- common/autobuild_common.sh@502 -- $ get_config_params 00:03:10.296 15:09:00 -- common/autotest_common.sh@409 -- $ xtrace_disable 00:03:10.296 15:09:00 -- common/autotest_common.sh@10 -- $ set +x 00:03:10.296 15:09:00 -- common/autobuild_common.sh@502 -- $ config_params='--enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --enable-ubsan --enable-asan --enable-coverage --with-ublk --with-raid5f --with-dpdk=/home/vagrant/spdk_repo/dpdk/build' 00:03:10.296 15:09:00 -- common/autobuild_common.sh@504 -- $ start_monitor_resources 00:03:10.296 15:09:00 -- pm/common@17 -- $ local monitor 00:03:10.296 15:09:00 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:03:10.296 15:09:00 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:03:10.296 15:09:00 -- pm/common@25 -- $ sleep 1 00:03:10.296 15:09:00 -- pm/common@21 -- $ date +%s 00:03:10.296 15:09:00 -- pm/common@21 -- $ date +%s 00:03:10.296 15:09:00 -- pm/common@21 -- $ /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-vmstat -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autobuild.sh.1732028940 00:03:10.296 15:09:00 -- pm/common@21 -- $ /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-cpu-load -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autobuild.sh.1732028940 00:03:10.296 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autobuild.sh.1732028940_collect-vmstat.pm.log 00:03:10.296 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autobuild.sh.1732028940_collect-cpu-load.pm.log 00:03:11.234 15:09:01 -- common/autobuild_common.sh@505 -- $ trap stop_monitor_resources EXIT 00:03:11.234 15:09:01 -- spdk/autobuild.sh@11 -- $ SPDK_TEST_AUTOBUILD= 00:03:11.234 15:09:01 -- spdk/autobuild.sh@12 -- $ umask 022 00:03:11.234 15:09:01 -- spdk/autobuild.sh@13 -- $ cd /home/vagrant/spdk_repo/spdk 00:03:11.234 15:09:01 -- spdk/autobuild.sh@16 -- $ date -u 00:03:11.234 Tue Nov 19 03:09:01 PM UTC 2024 00:03:11.234 15:09:01 -- spdk/autobuild.sh@17 -- $ git describe --tags 00:03:11.234 v25.01-pre-197-gdcc2ca8f3 00:03:11.234 15:09:01 -- spdk/autobuild.sh@19 -- $ '[' 1 -eq 1 ']' 00:03:11.234 15:09:01 -- spdk/autobuild.sh@20 -- $ run_test asan echo 'using asan' 00:03:11.234 15:09:01 -- common/autotest_common.sh@1105 -- $ '[' 3 -le 1 ']' 00:03:11.234 15:09:01 -- common/autotest_common.sh@1111 -- $ xtrace_disable 00:03:11.234 15:09:01 -- common/autotest_common.sh@10 -- $ set +x 00:03:11.234 ************************************ 00:03:11.234 START TEST asan 00:03:11.234 ************************************ 00:03:11.234 using asan 00:03:11.234 15:09:01 asan -- common/autotest_common.sh@1129 -- $ echo 'using asan' 00:03:11.234 00:03:11.234 real 0m0.001s 00:03:11.234 user 0m0.000s 00:03:11.234 sys 0m0.001s 00:03:11.234 15:09:01 asan -- common/autotest_common.sh@1130 -- $ xtrace_disable 00:03:11.234 15:09:01 asan -- common/autotest_common.sh@10 -- $ set +x 00:03:11.234 ************************************ 00:03:11.234 END TEST asan 00:03:11.234 ************************************ 00:03:11.234 15:09:01 -- spdk/autobuild.sh@23 -- $ '[' 1 -eq 1 ']' 00:03:11.234 15:09:01 -- spdk/autobuild.sh@24 -- $ run_test ubsan echo 'using ubsan' 00:03:11.234 15:09:01 -- common/autotest_common.sh@1105 -- $ '[' 3 -le 1 ']' 00:03:11.234 15:09:01 -- common/autotest_common.sh@1111 -- $ xtrace_disable 00:03:11.234 15:09:01 -- common/autotest_common.sh@10 -- $ set +x 00:03:11.234 ************************************ 00:03:11.234 START TEST ubsan 00:03:11.234 ************************************ 00:03:11.234 using ubsan 00:03:11.234 15:09:01 ubsan -- common/autotest_common.sh@1129 -- $ echo 'using ubsan' 00:03:11.234 00:03:11.234 real 0m0.001s 00:03:11.234 user 0m0.001s 00:03:11.234 sys 0m0.000s 00:03:11.234 15:09:01 ubsan -- common/autotest_common.sh@1130 -- $ xtrace_disable 00:03:11.234 15:09:01 ubsan -- common/autotest_common.sh@10 -- $ set +x 00:03:11.234 ************************************ 00:03:11.234 END TEST ubsan 00:03:11.234 ************************************ 00:03:11.494 15:09:01 -- spdk/autobuild.sh@27 -- $ '[' -n v22.11.4 ']' 00:03:11.494 15:09:01 -- spdk/autobuild.sh@28 -- $ build_native_dpdk 00:03:11.494 15:09:01 -- common/autobuild_common.sh@442 -- $ run_test build_native_dpdk _build_native_dpdk 00:03:11.494 15:09:01 -- common/autotest_common.sh@1105 -- $ '[' 2 -le 1 ']' 00:03:11.494 15:09:01 -- common/autotest_common.sh@1111 -- $ xtrace_disable 00:03:11.494 15:09:01 -- common/autotest_common.sh@10 -- $ set +x 00:03:11.494 ************************************ 00:03:11.494 START TEST build_native_dpdk 00:03:11.494 ************************************ 00:03:11.494 15:09:01 build_native_dpdk -- common/autotest_common.sh@1129 -- $ _build_native_dpdk 00:03:11.494 15:09:01 build_native_dpdk -- common/autobuild_common.sh@48 -- $ local external_dpdk_dir 00:03:11.494 15:09:01 build_native_dpdk -- common/autobuild_common.sh@49 -- $ local external_dpdk_base_dir 00:03:11.494 15:09:01 build_native_dpdk -- common/autobuild_common.sh@50 -- $ local compiler_version 00:03:11.494 15:09:01 build_native_dpdk -- common/autobuild_common.sh@51 -- $ local compiler 00:03:11.494 15:09:01 build_native_dpdk -- common/autobuild_common.sh@52 -- $ local dpdk_kmods 00:03:11.494 15:09:01 build_native_dpdk -- common/autobuild_common.sh@53 -- $ local repo=dpdk 00:03:11.494 15:09:01 build_native_dpdk -- common/autobuild_common.sh@55 -- $ compiler=gcc 00:03:11.494 15:09:01 build_native_dpdk -- common/autobuild_common.sh@61 -- $ export CC=gcc 00:03:11.494 15:09:01 build_native_dpdk -- common/autobuild_common.sh@61 -- $ CC=gcc 00:03:11.494 15:09:01 build_native_dpdk -- common/autobuild_common.sh@63 -- $ [[ gcc != *clang* ]] 00:03:11.494 15:09:01 build_native_dpdk -- common/autobuild_common.sh@63 -- $ [[ gcc != *gcc* ]] 00:03:11.494 15:09:01 build_native_dpdk -- common/autobuild_common.sh@68 -- $ gcc -dumpversion 00:03:11.494 15:09:01 build_native_dpdk -- common/autobuild_common.sh@68 -- $ compiler_version=13 00:03:11.494 15:09:01 build_native_dpdk -- common/autobuild_common.sh@69 -- $ compiler_version=13 00:03:11.494 15:09:01 build_native_dpdk -- common/autobuild_common.sh@70 -- $ external_dpdk_dir=/home/vagrant/spdk_repo/dpdk/build 00:03:11.494 15:09:01 build_native_dpdk -- common/autobuild_common.sh@71 -- $ dirname /home/vagrant/spdk_repo/dpdk/build 00:03:11.494 15:09:01 build_native_dpdk -- common/autobuild_common.sh@71 -- $ external_dpdk_base_dir=/home/vagrant/spdk_repo/dpdk 00:03:11.494 15:09:01 build_native_dpdk -- common/autobuild_common.sh@73 -- $ [[ ! -d /home/vagrant/spdk_repo/dpdk ]] 00:03:11.494 15:09:01 build_native_dpdk -- common/autobuild_common.sh@82 -- $ orgdir=/home/vagrant/spdk_repo/spdk 00:03:11.494 15:09:01 build_native_dpdk -- common/autobuild_common.sh@83 -- $ git -C /home/vagrant/spdk_repo/dpdk log --oneline -n 5 00:03:11.494 caf0f5d395 version: 22.11.4 00:03:11.494 7d6f1cc05f Revert "net/iavf: fix abnormal disable HW interrupt" 00:03:11.494 dc9c799c7d vhost: fix missing spinlock unlock 00:03:11.494 4307659a90 net/mlx5: fix LACP redirection in Rx domain 00:03:11.494 6ef77f2a5e net/gve: fix RX buffer size alignment 00:03:11.494 15:09:01 build_native_dpdk -- common/autobuild_common.sh@85 -- $ dpdk_cflags='-fPIC -g -fcommon' 00:03:11.494 15:09:01 build_native_dpdk -- common/autobuild_common.sh@86 -- $ dpdk_ldflags= 00:03:11.494 15:09:01 build_native_dpdk -- common/autobuild_common.sh@87 -- $ dpdk_ver=22.11.4 00:03:11.494 15:09:01 build_native_dpdk -- common/autobuild_common.sh@89 -- $ [[ gcc == *gcc* ]] 00:03:11.494 15:09:01 build_native_dpdk -- common/autobuild_common.sh@89 -- $ [[ 13 -ge 5 ]] 00:03:11.494 15:09:01 build_native_dpdk -- common/autobuild_common.sh@90 -- $ dpdk_cflags+=' -Werror' 00:03:11.494 15:09:01 build_native_dpdk -- common/autobuild_common.sh@93 -- $ [[ gcc == *gcc* ]] 00:03:11.494 15:09:01 build_native_dpdk -- common/autobuild_common.sh@93 -- $ [[ 13 -ge 10 ]] 00:03:11.494 15:09:01 build_native_dpdk -- common/autobuild_common.sh@94 -- $ dpdk_cflags+=' -Wno-stringop-overflow' 00:03:11.494 15:09:01 build_native_dpdk -- common/autobuild_common.sh@100 -- $ DPDK_DRIVERS=("bus" "bus/pci" "bus/vdev" "mempool/ring" "net/i40e" "net/i40e/base") 00:03:11.494 15:09:01 build_native_dpdk -- common/autobuild_common.sh@102 -- $ local mlx5_libs_added=n 00:03:11.494 15:09:01 build_native_dpdk -- common/autobuild_common.sh@103 -- $ [[ 0 -eq 1 ]] 00:03:11.494 15:09:01 build_native_dpdk -- common/autobuild_common.sh@103 -- $ [[ 0 -eq 1 ]] 00:03:11.494 15:09:01 build_native_dpdk -- common/autobuild_common.sh@139 -- $ [[ 0 -eq 1 ]] 00:03:11.494 15:09:01 build_native_dpdk -- common/autobuild_common.sh@167 -- $ cd /home/vagrant/spdk_repo/dpdk 00:03:11.494 15:09:01 build_native_dpdk -- common/autobuild_common.sh@168 -- $ uname -s 00:03:11.494 15:09:01 build_native_dpdk -- common/autobuild_common.sh@168 -- $ '[' Linux = Linux ']' 00:03:11.494 15:09:01 build_native_dpdk -- common/autobuild_common.sh@169 -- $ lt 22.11.4 21.11.0 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@373 -- $ cmp_versions 22.11.4 '<' 21.11.0 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@333 -- $ local ver1 ver1_l 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@334 -- $ local ver2 ver2_l 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@336 -- $ IFS=.-: 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@336 -- $ read -ra ver1 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@337 -- $ IFS=.-: 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@337 -- $ read -ra ver2 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@338 -- $ local 'op=<' 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@340 -- $ ver1_l=3 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@341 -- $ ver2_l=3 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@343 -- $ local lt=0 gt=0 eq=0 v 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@344 -- $ case "$op" in 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@345 -- $ : 1 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@364 -- $ (( v = 0 )) 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@364 -- $ (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@365 -- $ decimal 22 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@353 -- $ local d=22 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 22 =~ ^[0-9]+$ ]] 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@355 -- $ echo 22 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@365 -- $ ver1[v]=22 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@366 -- $ decimal 21 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@353 -- $ local d=21 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 21 =~ ^[0-9]+$ ]] 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@355 -- $ echo 21 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@366 -- $ ver2[v]=21 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@367 -- $ (( ver1[v] > ver2[v] )) 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@367 -- $ return 1 00:03:11.494 15:09:01 build_native_dpdk -- common/autobuild_common.sh@173 -- $ patch -p1 00:03:11.494 patching file config/rte_config.h 00:03:11.494 Hunk #1 succeeded at 60 (offset 1 line). 00:03:11.494 15:09:01 build_native_dpdk -- common/autobuild_common.sh@176 -- $ lt 22.11.4 24.07.0 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@373 -- $ cmp_versions 22.11.4 '<' 24.07.0 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@333 -- $ local ver1 ver1_l 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@334 -- $ local ver2 ver2_l 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@336 -- $ IFS=.-: 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@336 -- $ read -ra ver1 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@337 -- $ IFS=.-: 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@337 -- $ read -ra ver2 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@338 -- $ local 'op=<' 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@340 -- $ ver1_l=3 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@341 -- $ ver2_l=3 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@343 -- $ local lt=0 gt=0 eq=0 v 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@344 -- $ case "$op" in 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@345 -- $ : 1 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@364 -- $ (( v = 0 )) 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@364 -- $ (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@365 -- $ decimal 22 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@353 -- $ local d=22 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 22 =~ ^[0-9]+$ ]] 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@355 -- $ echo 22 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@365 -- $ ver1[v]=22 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@366 -- $ decimal 24 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@353 -- $ local d=24 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 24 =~ ^[0-9]+$ ]] 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@355 -- $ echo 24 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@366 -- $ ver2[v]=24 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@367 -- $ (( ver1[v] > ver2[v] )) 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@368 -- $ (( ver1[v] < ver2[v] )) 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@368 -- $ return 0 00:03:11.494 15:09:01 build_native_dpdk -- common/autobuild_common.sh@177 -- $ patch -p1 00:03:11.494 patching file lib/pcapng/rte_pcapng.c 00:03:11.494 Hunk #1 succeeded at 110 (offset -18 lines). 00:03:11.494 15:09:01 build_native_dpdk -- common/autobuild_common.sh@179 -- $ ge 22.11.4 24.07.0 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@376 -- $ cmp_versions 22.11.4 '>=' 24.07.0 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@333 -- $ local ver1 ver1_l 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@334 -- $ local ver2 ver2_l 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@336 -- $ IFS=.-: 00:03:11.494 15:09:01 build_native_dpdk -- scripts/common.sh@336 -- $ read -ra ver1 00:03:11.495 15:09:01 build_native_dpdk -- scripts/common.sh@337 -- $ IFS=.-: 00:03:11.495 15:09:01 build_native_dpdk -- scripts/common.sh@337 -- $ read -ra ver2 00:03:11.495 15:09:01 build_native_dpdk -- scripts/common.sh@338 -- $ local 'op=>=' 00:03:11.495 15:09:01 build_native_dpdk -- scripts/common.sh@340 -- $ ver1_l=3 00:03:11.495 15:09:01 build_native_dpdk -- scripts/common.sh@341 -- $ ver2_l=3 00:03:11.495 15:09:01 build_native_dpdk -- scripts/common.sh@343 -- $ local lt=0 gt=0 eq=0 v 00:03:11.495 15:09:01 build_native_dpdk -- scripts/common.sh@344 -- $ case "$op" in 00:03:11.495 15:09:01 build_native_dpdk -- scripts/common.sh@348 -- $ : 1 00:03:11.495 15:09:01 build_native_dpdk -- scripts/common.sh@364 -- $ (( v = 0 )) 00:03:11.495 15:09:01 build_native_dpdk -- scripts/common.sh@364 -- $ (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:03:11.495 15:09:01 build_native_dpdk -- scripts/common.sh@365 -- $ decimal 22 00:03:11.495 15:09:01 build_native_dpdk -- scripts/common.sh@353 -- $ local d=22 00:03:11.495 15:09:01 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 22 =~ ^[0-9]+$ ]] 00:03:11.495 15:09:01 build_native_dpdk -- scripts/common.sh@355 -- $ echo 22 00:03:11.495 15:09:01 build_native_dpdk -- scripts/common.sh@365 -- $ ver1[v]=22 00:03:11.495 15:09:01 build_native_dpdk -- scripts/common.sh@366 -- $ decimal 24 00:03:11.495 15:09:01 build_native_dpdk -- scripts/common.sh@353 -- $ local d=24 00:03:11.495 15:09:01 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 24 =~ ^[0-9]+$ ]] 00:03:11.495 15:09:01 build_native_dpdk -- scripts/common.sh@355 -- $ echo 24 00:03:11.495 15:09:01 build_native_dpdk -- scripts/common.sh@366 -- $ ver2[v]=24 00:03:11.495 15:09:01 build_native_dpdk -- scripts/common.sh@367 -- $ (( ver1[v] > ver2[v] )) 00:03:11.495 15:09:01 build_native_dpdk -- scripts/common.sh@368 -- $ (( ver1[v] < ver2[v] )) 00:03:11.495 15:09:01 build_native_dpdk -- scripts/common.sh@368 -- $ return 1 00:03:11.495 15:09:01 build_native_dpdk -- common/autobuild_common.sh@183 -- $ dpdk_kmods=false 00:03:11.495 15:09:01 build_native_dpdk -- common/autobuild_common.sh@184 -- $ uname -s 00:03:11.495 15:09:01 build_native_dpdk -- common/autobuild_common.sh@184 -- $ '[' Linux = FreeBSD ']' 00:03:11.495 15:09:01 build_native_dpdk -- common/autobuild_common.sh@188 -- $ printf %s, bus bus/pci bus/vdev mempool/ring net/i40e net/i40e/base 00:03:11.495 15:09:01 build_native_dpdk -- common/autobuild_common.sh@188 -- $ meson build-tmp --prefix=/home/vagrant/spdk_repo/dpdk/build --libdir lib -Denable_docs=false -Denable_kmods=false -Dtests=false -Dc_link_args= '-Dc_args=-fPIC -g -fcommon -Werror -Wno-stringop-overflow' -Dmachine=native -Denable_drivers=bus,bus/pci,bus/vdev,mempool/ring,net/i40e,net/i40e/base, 00:03:18.055 The Meson build system 00:03:18.055 Version: 1.5.0 00:03:18.055 Source dir: /home/vagrant/spdk_repo/dpdk 00:03:18.055 Build dir: /home/vagrant/spdk_repo/dpdk/build-tmp 00:03:18.055 Build type: native build 00:03:18.055 Program cat found: YES (/usr/bin/cat) 00:03:18.055 Project name: DPDK 00:03:18.055 Project version: 22.11.4 00:03:18.055 C compiler for the host machine: gcc (gcc 13.3.1 "gcc (GCC) 13.3.1 20240522 (Red Hat 13.3.1-1)") 00:03:18.055 C linker for the host machine: gcc ld.bfd 2.40-14 00:03:18.055 Host machine cpu family: x86_64 00:03:18.055 Host machine cpu: x86_64 00:03:18.055 Message: ## Building in Developer Mode ## 00:03:18.055 Program pkg-config found: YES (/usr/bin/pkg-config) 00:03:18.055 Program check-symbols.sh found: YES (/home/vagrant/spdk_repo/dpdk/buildtools/check-symbols.sh) 00:03:18.055 Program options-ibverbs-static.sh found: YES (/home/vagrant/spdk_repo/dpdk/buildtools/options-ibverbs-static.sh) 00:03:18.055 Program objdump found: YES (/usr/bin/objdump) 00:03:18.055 Program python3 found: YES (/usr/bin/python3) 00:03:18.055 Program cat found: YES (/usr/bin/cat) 00:03:18.055 config/meson.build:83: WARNING: The "machine" option is deprecated. Please use "cpu_instruction_set" instead. 00:03:18.055 Checking for size of "void *" : 8 00:03:18.055 Checking for size of "void *" : 8 (cached) 00:03:18.055 Library m found: YES 00:03:18.055 Library numa found: YES 00:03:18.055 Has header "numaif.h" : YES 00:03:18.055 Library fdt found: NO 00:03:18.055 Library execinfo found: NO 00:03:18.055 Has header "execinfo.h" : YES 00:03:18.055 Found pkg-config: YES (/usr/bin/pkg-config) 1.9.5 00:03:18.055 Run-time dependency libarchive found: NO (tried pkgconfig) 00:03:18.055 Run-time dependency libbsd found: NO (tried pkgconfig) 00:03:18.055 Run-time dependency jansson found: NO (tried pkgconfig) 00:03:18.055 Run-time dependency openssl found: YES 3.1.1 00:03:18.055 Run-time dependency libpcap found: YES 1.10.4 00:03:18.055 Has header "pcap.h" with dependency libpcap: YES 00:03:18.055 Compiler for C supports arguments -Wcast-qual: YES 00:03:18.055 Compiler for C supports arguments -Wdeprecated: YES 00:03:18.055 Compiler for C supports arguments -Wformat: YES 00:03:18.055 Compiler for C supports arguments -Wformat-nonliteral: NO 00:03:18.055 Compiler for C supports arguments -Wformat-security: NO 00:03:18.055 Compiler for C supports arguments -Wmissing-declarations: YES 00:03:18.055 Compiler for C supports arguments -Wmissing-prototypes: YES 00:03:18.055 Compiler for C supports arguments -Wnested-externs: YES 00:03:18.055 Compiler for C supports arguments -Wold-style-definition: YES 00:03:18.055 Compiler for C supports arguments -Wpointer-arith: YES 00:03:18.055 Compiler for C supports arguments -Wsign-compare: YES 00:03:18.055 Compiler for C supports arguments -Wstrict-prototypes: YES 00:03:18.055 Compiler for C supports arguments -Wundef: YES 00:03:18.055 Compiler for C supports arguments -Wwrite-strings: YES 00:03:18.055 Compiler for C supports arguments -Wno-address-of-packed-member: YES 00:03:18.055 Compiler for C supports arguments -Wno-packed-not-aligned: YES 00:03:18.055 Compiler for C supports arguments -Wno-missing-field-initializers: YES 00:03:18.055 Compiler for C supports arguments -Wno-zero-length-bounds: YES 00:03:18.055 Compiler for C supports arguments -mavx512f: YES 00:03:18.055 Checking if "AVX512 checking" compiles: YES 00:03:18.055 Fetching value of define "__SSE4_2__" : 1 00:03:18.055 Fetching value of define "__AES__" : 1 00:03:18.055 Fetching value of define "__AVX__" : 1 00:03:18.055 Fetching value of define "__AVX2__" : 1 00:03:18.055 Fetching value of define "__AVX512BW__" : 1 00:03:18.055 Fetching value of define "__AVX512CD__" : 1 00:03:18.055 Fetching value of define "__AVX512DQ__" : 1 00:03:18.055 Fetching value of define "__AVX512F__" : 1 00:03:18.055 Fetching value of define "__AVX512VL__" : 1 00:03:18.055 Fetching value of define "__PCLMUL__" : 1 00:03:18.055 Fetching value of define "__RDRND__" : 1 00:03:18.055 Fetching value of define "__RDSEED__" : 1 00:03:18.055 Fetching value of define "__VPCLMULQDQ__" : (undefined) 00:03:18.055 Compiler for C supports arguments -Wno-format-truncation: YES 00:03:18.055 Message: lib/kvargs: Defining dependency "kvargs" 00:03:18.055 Message: lib/telemetry: Defining dependency "telemetry" 00:03:18.055 Checking for function "getentropy" : YES 00:03:18.055 Message: lib/eal: Defining dependency "eal" 00:03:18.055 Message: lib/ring: Defining dependency "ring" 00:03:18.055 Message: lib/rcu: Defining dependency "rcu" 00:03:18.055 Message: lib/mempool: Defining dependency "mempool" 00:03:18.055 Message: lib/mbuf: Defining dependency "mbuf" 00:03:18.055 Fetching value of define "__PCLMUL__" : 1 (cached) 00:03:18.055 Fetching value of define "__AVX512F__" : 1 (cached) 00:03:18.055 Fetching value of define "__AVX512BW__" : 1 (cached) 00:03:18.055 Fetching value of define "__AVX512DQ__" : 1 (cached) 00:03:18.055 Fetching value of define "__AVX512VL__" : 1 (cached) 00:03:18.055 Fetching value of define "__VPCLMULQDQ__" : (undefined) (cached) 00:03:18.055 Compiler for C supports arguments -mpclmul: YES 00:03:18.055 Compiler for C supports arguments -maes: YES 00:03:18.055 Compiler for C supports arguments -mavx512f: YES (cached) 00:03:18.055 Compiler for C supports arguments -mavx512bw: YES 00:03:18.055 Compiler for C supports arguments -mavx512dq: YES 00:03:18.055 Compiler for C supports arguments -mavx512vl: YES 00:03:18.055 Compiler for C supports arguments -mvpclmulqdq: YES 00:03:18.055 Compiler for C supports arguments -mavx2: YES 00:03:18.055 Compiler for C supports arguments -mavx: YES 00:03:18.055 Message: lib/net: Defining dependency "net" 00:03:18.055 Message: lib/meter: Defining dependency "meter" 00:03:18.055 Message: lib/ethdev: Defining dependency "ethdev" 00:03:18.055 Message: lib/pci: Defining dependency "pci" 00:03:18.055 Message: lib/cmdline: Defining dependency "cmdline" 00:03:18.055 Message: lib/metrics: Defining dependency "metrics" 00:03:18.055 Message: lib/hash: Defining dependency "hash" 00:03:18.055 Message: lib/timer: Defining dependency "timer" 00:03:18.055 Fetching value of define "__AVX2__" : 1 (cached) 00:03:18.055 Fetching value of define "__AVX512F__" : 1 (cached) 00:03:18.055 Fetching value of define "__AVX512VL__" : 1 (cached) 00:03:18.055 Fetching value of define "__AVX512CD__" : 1 (cached) 00:03:18.055 Fetching value of define "__AVX512BW__" : 1 (cached) 00:03:18.055 Message: lib/acl: Defining dependency "acl" 00:03:18.055 Message: lib/bbdev: Defining dependency "bbdev" 00:03:18.055 Message: lib/bitratestats: Defining dependency "bitratestats" 00:03:18.055 Run-time dependency libelf found: YES 0.191 00:03:18.055 Message: lib/bpf: Defining dependency "bpf" 00:03:18.055 Message: lib/cfgfile: Defining dependency "cfgfile" 00:03:18.056 Message: lib/compressdev: Defining dependency "compressdev" 00:03:18.056 Message: lib/cryptodev: Defining dependency "cryptodev" 00:03:18.056 Message: lib/distributor: Defining dependency "distributor" 00:03:18.056 Message: lib/efd: Defining dependency "efd" 00:03:18.056 Message: lib/eventdev: Defining dependency "eventdev" 00:03:18.056 Message: lib/gpudev: Defining dependency "gpudev" 00:03:18.056 Message: lib/gro: Defining dependency "gro" 00:03:18.056 Message: lib/gso: Defining dependency "gso" 00:03:18.056 Message: lib/ip_frag: Defining dependency "ip_frag" 00:03:18.056 Message: lib/jobstats: Defining dependency "jobstats" 00:03:18.056 Message: lib/latencystats: Defining dependency "latencystats" 00:03:18.056 Message: lib/lpm: Defining dependency "lpm" 00:03:18.056 Fetching value of define "__AVX512F__" : 1 (cached) 00:03:18.056 Fetching value of define "__AVX512DQ__" : 1 (cached) 00:03:18.056 Fetching value of define "__AVX512IFMA__" : (undefined) 00:03:18.056 Compiler for C supports arguments -mavx512f -mavx512dq -mavx512ifma: YES 00:03:18.056 Message: lib/member: Defining dependency "member" 00:03:18.056 Message: lib/pcapng: Defining dependency "pcapng" 00:03:18.056 Compiler for C supports arguments -Wno-cast-qual: YES 00:03:18.056 Message: lib/power: Defining dependency "power" 00:03:18.056 Message: lib/rawdev: Defining dependency "rawdev" 00:03:18.056 Message: lib/regexdev: Defining dependency "regexdev" 00:03:18.056 Message: lib/dmadev: Defining dependency "dmadev" 00:03:18.056 Message: lib/rib: Defining dependency "rib" 00:03:18.056 Message: lib/reorder: Defining dependency "reorder" 00:03:18.056 Message: lib/sched: Defining dependency "sched" 00:03:18.056 Message: lib/security: Defining dependency "security" 00:03:18.056 Message: lib/stack: Defining dependency "stack" 00:03:18.056 Has header "linux/userfaultfd.h" : YES 00:03:18.056 Message: lib/vhost: Defining dependency "vhost" 00:03:18.056 Message: lib/ipsec: Defining dependency "ipsec" 00:03:18.056 Fetching value of define "__AVX512F__" : 1 (cached) 00:03:18.056 Fetching value of define "__AVX512DQ__" : 1 (cached) 00:03:18.056 Fetching value of define "__AVX512BW__" : 1 (cached) 00:03:18.056 Message: lib/fib: Defining dependency "fib" 00:03:18.056 Message: lib/port: Defining dependency "port" 00:03:18.056 Message: lib/pdump: Defining dependency "pdump" 00:03:18.056 Message: lib/table: Defining dependency "table" 00:03:18.056 Message: lib/pipeline: Defining dependency "pipeline" 00:03:18.056 Message: lib/graph: Defining dependency "graph" 00:03:18.056 Message: lib/node: Defining dependency "node" 00:03:18.056 Compiler for C supports arguments -Wno-format-truncation: YES (cached) 00:03:18.056 Message: drivers/bus/pci: Defining dependency "bus_pci" 00:03:18.056 Message: drivers/bus/vdev: Defining dependency "bus_vdev" 00:03:18.056 Message: drivers/mempool/ring: Defining dependency "mempool_ring" 00:03:18.056 Compiler for C supports arguments -Wno-sign-compare: YES 00:03:18.056 Compiler for C supports arguments -Wno-unused-value: YES 00:03:18.056 Compiler for C supports arguments -Wno-format: YES 00:03:18.056 Compiler for C supports arguments -Wno-format-security: YES 00:03:18.056 Compiler for C supports arguments -Wno-format-nonliteral: YES 00:03:18.056 Compiler for C supports arguments -Wno-strict-aliasing: YES 00:03:18.623 Compiler for C supports arguments -Wno-unused-but-set-variable: YES 00:03:18.623 Compiler for C supports arguments -Wno-unused-parameter: YES 00:03:18.623 Fetching value of define "__AVX2__" : 1 (cached) 00:03:18.623 Fetching value of define "__AVX512F__" : 1 (cached) 00:03:18.623 Fetching value of define "__AVX512BW__" : 1 (cached) 00:03:18.623 Compiler for C supports arguments -mavx512f: YES (cached) 00:03:18.623 Compiler for C supports arguments -mavx512bw: YES (cached) 00:03:18.623 Compiler for C supports arguments -march=skylake-avx512: YES 00:03:18.623 Message: drivers/net/i40e: Defining dependency "net_i40e" 00:03:18.623 Program doxygen found: YES (/usr/local/bin/doxygen) 00:03:18.623 Configuring doxy-api.conf using configuration 00:03:18.623 Program sphinx-build found: NO 00:03:18.623 Configuring rte_build_config.h using configuration 00:03:18.623 Message: 00:03:18.623 ================= 00:03:18.623 Applications Enabled 00:03:18.623 ================= 00:03:18.623 00:03:18.623 apps: 00:03:18.623 dumpcap, pdump, proc-info, test-acl, test-bbdev, test-cmdline, test-compress-perf, test-crypto-perf, 00:03:18.623 test-eventdev, test-fib, test-flow-perf, test-gpudev, test-pipeline, test-pmd, test-regex, test-sad, 00:03:18.623 test-security-perf, 00:03:18.623 00:03:18.623 Message: 00:03:18.623 ================= 00:03:18.623 Libraries Enabled 00:03:18.623 ================= 00:03:18.623 00:03:18.623 libs: 00:03:18.623 kvargs, telemetry, eal, ring, rcu, mempool, mbuf, net, 00:03:18.623 meter, ethdev, pci, cmdline, metrics, hash, timer, acl, 00:03:18.623 bbdev, bitratestats, bpf, cfgfile, compressdev, cryptodev, distributor, efd, 00:03:18.623 eventdev, gpudev, gro, gso, ip_frag, jobstats, latencystats, lpm, 00:03:18.623 member, pcapng, power, rawdev, regexdev, dmadev, rib, reorder, 00:03:18.623 sched, security, stack, vhost, ipsec, fib, port, pdump, 00:03:18.623 table, pipeline, graph, node, 00:03:18.623 00:03:18.623 Message: 00:03:18.623 =============== 00:03:18.623 Drivers Enabled 00:03:18.623 =============== 00:03:18.623 00:03:18.623 common: 00:03:18.623 00:03:18.623 bus: 00:03:18.623 pci, vdev, 00:03:18.623 mempool: 00:03:18.623 ring, 00:03:18.623 dma: 00:03:18.623 00:03:18.623 net: 00:03:18.623 i40e, 00:03:18.623 raw: 00:03:18.623 00:03:18.623 crypto: 00:03:18.623 00:03:18.623 compress: 00:03:18.623 00:03:18.623 regex: 00:03:18.623 00:03:18.623 vdpa: 00:03:18.623 00:03:18.623 event: 00:03:18.623 00:03:18.623 baseband: 00:03:18.623 00:03:18.623 gpu: 00:03:18.623 00:03:18.623 00:03:18.623 Message: 00:03:18.623 ================= 00:03:18.623 Content Skipped 00:03:18.623 ================= 00:03:18.623 00:03:18.623 apps: 00:03:18.623 00:03:18.623 libs: 00:03:18.623 kni: explicitly disabled via build config (deprecated lib) 00:03:18.623 flow_classify: explicitly disabled via build config (deprecated lib) 00:03:18.623 00:03:18.623 drivers: 00:03:18.623 common/cpt: not in enabled drivers build config 00:03:18.623 common/dpaax: not in enabled drivers build config 00:03:18.623 common/iavf: not in enabled drivers build config 00:03:18.623 common/idpf: not in enabled drivers build config 00:03:18.623 common/mvep: not in enabled drivers build config 00:03:18.623 common/octeontx: not in enabled drivers build config 00:03:18.623 bus/auxiliary: not in enabled drivers build config 00:03:18.623 bus/dpaa: not in enabled drivers build config 00:03:18.623 bus/fslmc: not in enabled drivers build config 00:03:18.623 bus/ifpga: not in enabled drivers build config 00:03:18.623 bus/vmbus: not in enabled drivers build config 00:03:18.623 common/cnxk: not in enabled drivers build config 00:03:18.623 common/mlx5: not in enabled drivers build config 00:03:18.623 common/qat: not in enabled drivers build config 00:03:18.623 common/sfc_efx: not in enabled drivers build config 00:03:18.623 mempool/bucket: not in enabled drivers build config 00:03:18.624 mempool/cnxk: not in enabled drivers build config 00:03:18.624 mempool/dpaa: not in enabled drivers build config 00:03:18.624 mempool/dpaa2: not in enabled drivers build config 00:03:18.624 mempool/octeontx: not in enabled drivers build config 00:03:18.624 mempool/stack: not in enabled drivers build config 00:03:18.624 dma/cnxk: not in enabled drivers build config 00:03:18.624 dma/dpaa: not in enabled drivers build config 00:03:18.624 dma/dpaa2: not in enabled drivers build config 00:03:18.624 dma/hisilicon: not in enabled drivers build config 00:03:18.624 dma/idxd: not in enabled drivers build config 00:03:18.624 dma/ioat: not in enabled drivers build config 00:03:18.624 dma/skeleton: not in enabled drivers build config 00:03:18.624 net/af_packet: not in enabled drivers build config 00:03:18.624 net/af_xdp: not in enabled drivers build config 00:03:18.624 net/ark: not in enabled drivers build config 00:03:18.624 net/atlantic: not in enabled drivers build config 00:03:18.624 net/avp: not in enabled drivers build config 00:03:18.624 net/axgbe: not in enabled drivers build config 00:03:18.624 net/bnx2x: not in enabled drivers build config 00:03:18.624 net/bnxt: not in enabled drivers build config 00:03:18.624 net/bonding: not in enabled drivers build config 00:03:18.624 net/cnxk: not in enabled drivers build config 00:03:18.624 net/cxgbe: not in enabled drivers build config 00:03:18.624 net/dpaa: not in enabled drivers build config 00:03:18.624 net/dpaa2: not in enabled drivers build config 00:03:18.624 net/e1000: not in enabled drivers build config 00:03:18.624 net/ena: not in enabled drivers build config 00:03:18.624 net/enetc: not in enabled drivers build config 00:03:18.624 net/enetfec: not in enabled drivers build config 00:03:18.624 net/enic: not in enabled drivers build config 00:03:18.624 net/failsafe: not in enabled drivers build config 00:03:18.624 net/fm10k: not in enabled drivers build config 00:03:18.624 net/gve: not in enabled drivers build config 00:03:18.624 net/hinic: not in enabled drivers build config 00:03:18.624 net/hns3: not in enabled drivers build config 00:03:18.624 net/iavf: not in enabled drivers build config 00:03:18.624 net/ice: not in enabled drivers build config 00:03:18.624 net/idpf: not in enabled drivers build config 00:03:18.624 net/igc: not in enabled drivers build config 00:03:18.624 net/ionic: not in enabled drivers build config 00:03:18.624 net/ipn3ke: not in enabled drivers build config 00:03:18.624 net/ixgbe: not in enabled drivers build config 00:03:18.624 net/kni: not in enabled drivers build config 00:03:18.624 net/liquidio: not in enabled drivers build config 00:03:18.624 net/mana: not in enabled drivers build config 00:03:18.624 net/memif: not in enabled drivers build config 00:03:18.624 net/mlx4: not in enabled drivers build config 00:03:18.624 net/mlx5: not in enabled drivers build config 00:03:18.624 net/mvneta: not in enabled drivers build config 00:03:18.624 net/mvpp2: not in enabled drivers build config 00:03:18.624 net/netvsc: not in enabled drivers build config 00:03:18.624 net/nfb: not in enabled drivers build config 00:03:18.624 net/nfp: not in enabled drivers build config 00:03:18.624 net/ngbe: not in enabled drivers build config 00:03:18.624 net/null: not in enabled drivers build config 00:03:18.624 net/octeontx: not in enabled drivers build config 00:03:18.624 net/octeon_ep: not in enabled drivers build config 00:03:18.624 net/pcap: not in enabled drivers build config 00:03:18.624 net/pfe: not in enabled drivers build config 00:03:18.624 net/qede: not in enabled drivers build config 00:03:18.624 net/ring: not in enabled drivers build config 00:03:18.624 net/sfc: not in enabled drivers build config 00:03:18.624 net/softnic: not in enabled drivers build config 00:03:18.624 net/tap: not in enabled drivers build config 00:03:18.624 net/thunderx: not in enabled drivers build config 00:03:18.624 net/txgbe: not in enabled drivers build config 00:03:18.624 net/vdev_netvsc: not in enabled drivers build config 00:03:18.624 net/vhost: not in enabled drivers build config 00:03:18.624 net/virtio: not in enabled drivers build config 00:03:18.624 net/vmxnet3: not in enabled drivers build config 00:03:18.624 raw/cnxk_bphy: not in enabled drivers build config 00:03:18.624 raw/cnxk_gpio: not in enabled drivers build config 00:03:18.624 raw/dpaa2_cmdif: not in enabled drivers build config 00:03:18.624 raw/ifpga: not in enabled drivers build config 00:03:18.624 raw/ntb: not in enabled drivers build config 00:03:18.624 raw/skeleton: not in enabled drivers build config 00:03:18.624 crypto/armv8: not in enabled drivers build config 00:03:18.624 crypto/bcmfs: not in enabled drivers build config 00:03:18.624 crypto/caam_jr: not in enabled drivers build config 00:03:18.624 crypto/ccp: not in enabled drivers build config 00:03:18.624 crypto/cnxk: not in enabled drivers build config 00:03:18.624 crypto/dpaa_sec: not in enabled drivers build config 00:03:18.624 crypto/dpaa2_sec: not in enabled drivers build config 00:03:18.624 crypto/ipsec_mb: not in enabled drivers build config 00:03:18.624 crypto/mlx5: not in enabled drivers build config 00:03:18.624 crypto/mvsam: not in enabled drivers build config 00:03:18.624 crypto/nitrox: not in enabled drivers build config 00:03:18.624 crypto/null: not in enabled drivers build config 00:03:18.624 crypto/octeontx: not in enabled drivers build config 00:03:18.624 crypto/openssl: not in enabled drivers build config 00:03:18.624 crypto/scheduler: not in enabled drivers build config 00:03:18.624 crypto/uadk: not in enabled drivers build config 00:03:18.624 crypto/virtio: not in enabled drivers build config 00:03:18.624 compress/isal: not in enabled drivers build config 00:03:18.624 compress/mlx5: not in enabled drivers build config 00:03:18.624 compress/octeontx: not in enabled drivers build config 00:03:18.624 compress/zlib: not in enabled drivers build config 00:03:18.624 regex/mlx5: not in enabled drivers build config 00:03:18.624 regex/cn9k: not in enabled drivers build config 00:03:18.624 vdpa/ifc: not in enabled drivers build config 00:03:18.624 vdpa/mlx5: not in enabled drivers build config 00:03:18.624 vdpa/sfc: not in enabled drivers build config 00:03:18.624 event/cnxk: not in enabled drivers build config 00:03:18.624 event/dlb2: not in enabled drivers build config 00:03:18.624 event/dpaa: not in enabled drivers build config 00:03:18.624 event/dpaa2: not in enabled drivers build config 00:03:18.624 event/dsw: not in enabled drivers build config 00:03:18.624 event/opdl: not in enabled drivers build config 00:03:18.624 event/skeleton: not in enabled drivers build config 00:03:18.624 event/sw: not in enabled drivers build config 00:03:18.624 event/octeontx: not in enabled drivers build config 00:03:18.624 baseband/acc: not in enabled drivers build config 00:03:18.624 baseband/fpga_5gnr_fec: not in enabled drivers build config 00:03:18.624 baseband/fpga_lte_fec: not in enabled drivers build config 00:03:18.624 baseband/la12xx: not in enabled drivers build config 00:03:18.624 baseband/null: not in enabled drivers build config 00:03:18.624 baseband/turbo_sw: not in enabled drivers build config 00:03:18.624 gpu/cuda: not in enabled drivers build config 00:03:18.624 00:03:18.624 00:03:18.624 Build targets in project: 311 00:03:18.624 00:03:18.624 DPDK 22.11.4 00:03:18.624 00:03:18.624 User defined options 00:03:18.624 libdir : lib 00:03:18.624 prefix : /home/vagrant/spdk_repo/dpdk/build 00:03:18.624 c_args : -fPIC -g -fcommon -Werror -Wno-stringop-overflow 00:03:18.624 c_link_args : 00:03:18.624 enable_docs : false 00:03:18.624 enable_drivers: bus,bus/pci,bus/vdev,mempool/ring,net/i40e,net/i40e/base, 00:03:18.624 enable_kmods : false 00:03:18.624 machine : native 00:03:18.624 tests : false 00:03:18.624 00:03:18.624 Found ninja-1.11.1.git.kitware.jobserver-1 at /usr/local/bin/ninja 00:03:18.624 WARNING: Running the setup command as `meson [options]` instead of `meson setup [options]` is ambiguous and deprecated. 00:03:18.883 15:09:09 build_native_dpdk -- common/autobuild_common.sh@192 -- $ ninja -C /home/vagrant/spdk_repo/dpdk/build-tmp -j10 00:03:18.883 ninja: Entering directory `/home/vagrant/spdk_repo/dpdk/build-tmp' 00:03:18.883 [1/740] Generating lib/rte_kvargs_def with a custom command 00:03:18.883 [2/740] Generating lib/rte_telemetry_def with a custom command 00:03:18.883 [3/740] Generating lib/rte_kvargs_mingw with a custom command 00:03:18.883 [4/740] Generating lib/rte_telemetry_mingw with a custom command 00:03:18.883 [5/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_class.c.o 00:03:18.883 [6/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_debug.c.o 00:03:18.883 [7/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_errno.c.o 00:03:18.883 [8/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_config.c.o 00:03:19.140 [9/740] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry_data.c.o 00:03:19.140 [10/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_bus.c.o 00:03:19.140 [11/740] Compiling C object lib/librte_kvargs.a.p/kvargs_rte_kvargs.c.o 00:03:19.140 [12/740] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry_legacy.c.o 00:03:19.140 [13/740] Linking static target lib/librte_kvargs.a 00:03:19.140 [14/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_hexdump.c.o 00:03:19.140 [15/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_launch.c.o 00:03:19.140 [16/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_devargs.c.o 00:03:19.140 [17/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_dev.c.o 00:03:19.140 [18/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_mcfg.c.o 00:03:19.140 [19/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_interrupts.c.o 00:03:19.140 [20/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_lcore.c.o 00:03:19.140 [21/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_log.c.o 00:03:19.140 [22/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_string_fns.c.o 00:03:19.397 [23/740] Generating lib/kvargs.sym_chk with a custom command (wrapped by meson to capture output) 00:03:19.397 [24/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memalloc.c.o 00:03:19.397 [25/740] Linking target lib/librte_kvargs.so.23.0 00:03:19.397 [26/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_timer.c.o 00:03:19.397 [27/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memzone.c.o 00:03:19.397 [28/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_tailqs.c.o 00:03:19.397 [29/740] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry.c.o 00:03:19.397 [30/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_fbarray.c.o 00:03:19.397 [31/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_thread.c.o 00:03:19.397 [32/740] Linking static target lib/librte_telemetry.a 00:03:19.397 [33/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_uuid.c.o 00:03:19.397 [34/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_points.c.o 00:03:19.397 [35/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memory.c.o 00:03:19.397 [36/740] Compiling C object lib/librte_eal.a.p/eal_common_rte_reciprocal.c.o 00:03:19.655 [37/740] Compiling C object lib/librte_eal.a.p/eal_common_rte_version.c.o 00:03:19.655 [38/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_cpuflags.c.o 00:03:19.655 [39/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_hypervisor.c.o 00:03:19.655 [40/740] Compiling C object lib/librte_eal.a.p/eal_common_malloc_elem.c.o 00:03:19.655 [41/740] Generating symbol file lib/librte_kvargs.so.23.0.p/librte_kvargs.so.23.0.symbols 00:03:19.655 [42/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_dynmem.c.o 00:03:19.655 [43/740] Generating lib/telemetry.sym_chk with a custom command (wrapped by meson to capture output) 00:03:19.655 [44/740] Linking target lib/librte_telemetry.so.23.0 00:03:19.912 [45/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_ctf.c.o 00:03:19.912 [46/740] Compiling C object lib/librte_eal.a.p/eal_common_malloc_heap.c.o 00:03:19.912 [47/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace.c.o 00:03:19.913 [48/740] Compiling C object lib/librte_eal.a.p/eal_common_rte_random.c.o 00:03:19.913 [49/740] Compiling C object lib/librte_eal.a.p/eal_common_rte_service.c.o 00:03:19.913 [50/740] Generating symbol file lib/librte_telemetry.so.23.0.p/librte_telemetry.so.23.0.symbols 00:03:19.913 [51/740] Compiling C object lib/librte_eal.a.p/eal_common_rte_keepalive.c.o 00:03:19.913 [52/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_utils.c.o 00:03:19.913 [53/740] Compiling C object lib/librte_eal.a.p/eal_unix_eal_debug.c.o 00:03:19.913 [54/740] Compiling C object lib/librte_eal.a.p/eal_common_hotplug_mp.c.o 00:03:19.913 [55/740] Compiling C object lib/librte_eal.a.p/eal_unix_eal_file.c.o 00:03:19.913 [56/740] Compiling C object lib/librte_eal.a.p/eal_unix_eal_firmware.c.o 00:03:19.913 [57/740] Compiling C object lib/librte_eal.a.p/eal_unix_eal_filesystem.c.o 00:03:19.913 [58/740] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_timer.c.o 00:03:19.913 [59/740] Compiling C object lib/librte_eal.a.p/eal_common_malloc_mp.c.o 00:03:19.913 [60/740] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_thread.c.o 00:03:19.913 [61/740] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_memory.c.o 00:03:19.913 [62/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_proc.c.o 00:03:20.172 [63/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_cpuflags.c.o 00:03:20.173 [64/740] Compiling C object lib/librte_eal.a.p/eal_unix_rte_thread.c.o 00:03:20.173 [65/740] Compiling C object lib/librte_eal.a.p/eal_common_rte_malloc.c.o 00:03:20.173 [66/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_log.c.o 00:03:20.173 [67/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_lcore.c.o 00:03:20.173 [68/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_alarm.c.o 00:03:20.173 [69/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_thread.c.o 00:03:20.173 [70/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_dev.c.o 00:03:20.173 [71/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_timer.c.o 00:03:20.173 [72/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_vfio_mp_sync.c.o 00:03:20.173 [73/740] Compiling C object lib/librte_eal.a.p/eal_x86_rte_cpuflags.c.o 00:03:20.173 [74/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_hugepage_info.c.o 00:03:20.173 [75/740] Compiling C object lib/librte_eal.a.p/eal_x86_rte_spinlock.c.o 00:03:20.173 [76/740] Compiling C object lib/librte_eal.a.p/eal_x86_rte_hypervisor.c.o 00:03:20.173 [77/740] Generating lib/rte_eal_def with a custom command 00:03:20.173 [78/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_options.c.o 00:03:20.173 [79/740] Generating lib/rte_eal_mingw with a custom command 00:03:20.431 [80/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal.c.o 00:03:20.431 [81/740] Generating lib/rte_ring_def with a custom command 00:03:20.431 [82/740] Generating lib/rte_ring_mingw with a custom command 00:03:20.431 [83/740] Compiling C object lib/librte_eal.a.p/eal_x86_rte_cycles.c.o 00:03:20.431 [84/740] Generating lib/rte_rcu_mingw with a custom command 00:03:20.431 [85/740] Generating lib/rte_rcu_def with a custom command 00:03:20.431 [86/740] Compiling C object lib/librte_eal.a.p/eal_x86_rte_power_intrinsics.c.o 00:03:20.431 [87/740] Compiling C object lib/librte_ring.a.p/ring_rte_ring.c.o 00:03:20.431 [88/740] Linking static target lib/librte_ring.a 00:03:20.431 [89/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_interrupts.c.o 00:03:20.431 [90/740] Generating lib/rte_mempool_def with a custom command 00:03:20.431 [91/740] Generating lib/rte_mempool_mingw with a custom command 00:03:20.431 [92/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_memalloc.c.o 00:03:20.690 [93/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_memory.c.o 00:03:20.690 [94/740] Generating lib/ring.sym_chk with a custom command (wrapped by meson to capture output) 00:03:20.690 [95/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_vfio.c.o 00:03:20.690 [96/740] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool_ops_default.c.o 00:03:20.690 [97/740] Linking static target lib/librte_eal.a 00:03:20.690 [98/740] Compiling C object lib/librte_mempool.a.p/mempool_mempool_trace_points.c.o 00:03:20.690 [99/740] Generating lib/rte_mbuf_def with a custom command 00:03:20.690 [100/740] Generating lib/rte_mbuf_mingw with a custom command 00:03:20.690 [101/740] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool_ops.c.o 00:03:20.957 [102/740] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_ptype.c.o 00:03:20.957 [103/740] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_pool_ops.c.o 00:03:20.957 [104/740] Compiling C object lib/librte_rcu.a.p/rcu_rte_rcu_qsbr.c.o 00:03:20.957 [105/740] Linking static target lib/librte_rcu.a 00:03:21.230 [106/740] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool.c.o 00:03:21.230 [107/740] Linking static target lib/librte_mempool.a 00:03:21.230 [108/740] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_dyn.c.o 00:03:21.230 [109/740] Compiling C object lib/librte_net.a.p/net_rte_ether.c.o 00:03:21.230 [110/740] Generating lib/rte_net_def with a custom command 00:03:21.230 [111/740] Compiling C object lib/net/libnet_crc_avx512_lib.a.p/net_crc_avx512.c.o 00:03:21.230 [112/740] Linking static target lib/net/libnet_crc_avx512_lib.a 00:03:21.230 [113/740] Compiling C object lib/librte_net.a.p/net_rte_arp.c.o 00:03:21.230 [114/740] Generating lib/rte_net_mingw with a custom command 00:03:21.230 [115/740] Generating lib/rte_meter_def with a custom command 00:03:21.230 [116/740] Generating lib/rte_meter_mingw with a custom command 00:03:21.230 [117/740] Compiling C object lib/librte_net.a.p/net_rte_net_crc.c.o 00:03:21.230 [118/740] Generating lib/rcu.sym_chk with a custom command (wrapped by meson to capture output) 00:03:21.488 [119/740] Compiling C object lib/librte_net.a.p/net_rte_net.c.o 00:03:21.488 [120/740] Compiling C object lib/librte_meter.a.p/meter_rte_meter.c.o 00:03:21.488 [121/740] Linking static target lib/librte_meter.a 00:03:21.488 [122/740] Compiling C object lib/librte_net.a.p/net_net_crc_sse.c.o 00:03:21.488 [123/740] Linking static target lib/librte_net.a 00:03:21.488 [124/740] Generating lib/meter.sym_chk with a custom command (wrapped by meson to capture output) 00:03:21.747 [125/740] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf.c.o 00:03:21.747 [126/740] Linking static target lib/librte_mbuf.a 00:03:21.747 [127/740] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_profile.c.o 00:03:21.747 [128/740] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_trace_points.c.o 00:03:21.747 [129/740] Generating lib/net.sym_chk with a custom command (wrapped by meson to capture output) 00:03:21.747 [130/740] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_private.c.o 00:03:21.747 [131/740] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_class_eth.c.o 00:03:21.747 [132/740] Generating lib/mempool.sym_chk with a custom command (wrapped by meson to capture output) 00:03:21.747 [133/740] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_driver.c.o 00:03:22.006 [134/740] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev_cman.c.o 00:03:22.264 [135/740] Generating lib/mbuf.sym_chk with a custom command (wrapped by meson to capture output) 00:03:22.264 [136/740] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_telemetry.c.o 00:03:22.264 [137/740] Generating lib/rte_ethdev_def with a custom command 00:03:22.264 [138/740] Generating lib/rte_ethdev_mingw with a custom command 00:03:22.264 [139/740] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_mtr.c.o 00:03:22.264 [140/740] Generating lib/rte_pci_def with a custom command 00:03:22.264 [141/740] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8079.c.o 00:03:22.264 [142/740] Generating lib/rte_pci_mingw with a custom command 00:03:22.264 [143/740] Compiling C object lib/librte_pci.a.p/pci_rte_pci.c.o 00:03:22.264 [144/740] Linking static target lib/librte_pci.a 00:03:22.523 [145/740] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_common.c.o 00:03:22.523 [146/740] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_tm.c.o 00:03:22.523 [147/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline.c.o 00:03:22.523 [148/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_cirbuf.c.o 00:03:22.523 [149/740] Generating lib/pci.sym_chk with a custom command (wrapped by meson to capture output) 00:03:22.523 [150/740] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8472.c.o 00:03:22.523 [151/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse.c.o 00:03:22.781 [152/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_portlist.c.o 00:03:22.781 [153/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_num.c.o 00:03:22.781 [154/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_string.c.o 00:03:22.781 [155/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_socket.c.o 00:03:22.781 [156/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_vt100.c.o 00:03:22.781 [157/740] Generating lib/rte_cmdline_def with a custom command 00:03:22.781 [158/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_os_unix.c.o 00:03:22.781 [159/740] Generating lib/rte_cmdline_mingw with a custom command 00:03:22.781 [160/740] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8636.c.o 00:03:22.781 [161/740] Generating lib/rte_metrics_def with a custom command 00:03:22.781 [162/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_etheraddr.c.o 00:03:22.781 [163/740] Generating lib/rte_metrics_mingw with a custom command 00:03:22.781 [164/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_rdline.c.o 00:03:22.781 [165/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_ipaddr.c.o 00:03:22.781 [166/740] Linking static target lib/librte_cmdline.a 00:03:22.781 [167/740] Generating lib/rte_hash_def with a custom command 00:03:22.781 [168/740] Compiling C object lib/librte_metrics.a.p/metrics_rte_metrics.c.o 00:03:22.781 [169/740] Generating lib/rte_hash_mingw with a custom command 00:03:23.039 [170/740] Generating lib/rte_timer_def with a custom command 00:03:23.039 [171/740] Compiling C object lib/librte_hash.a.p/hash_rte_fbk_hash.c.o 00:03:23.039 [172/740] Generating lib/rte_timer_mingw with a custom command 00:03:23.039 [173/740] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_flow.c.o 00:03:23.039 [174/740] Compiling C object lib/librte_metrics.a.p/metrics_rte_metrics_telemetry.c.o 00:03:23.039 [175/740] Linking static target lib/librte_metrics.a 00:03:23.297 [176/740] Compiling C object lib/librte_timer.a.p/timer_rte_timer.c.o 00:03:23.297 [177/740] Linking static target lib/librte_timer.a 00:03:23.556 [178/740] Generating lib/metrics.sym_chk with a custom command (wrapped by meson to capture output) 00:03:23.556 [179/740] Compiling C object lib/librte_acl.a.p/acl_acl_gen.c.o 00:03:23.556 [180/740] Compiling C object lib/librte_hash.a.p/hash_rte_thash.c.o 00:03:23.556 [181/740] Generating lib/timer.sym_chk with a custom command (wrapped by meson to capture output) 00:03:23.814 [182/740] Compiling C object lib/librte_acl.a.p/acl_acl_run_scalar.c.o 00:03:23.814 [183/740] Generating lib/cmdline.sym_chk with a custom command (wrapped by meson to capture output) 00:03:23.814 [184/740] Generating lib/rte_acl_def with a custom command 00:03:23.814 [185/740] Generating lib/rte_acl_mingw with a custom command 00:03:23.814 [186/740] Compiling C object lib/librte_acl.a.p/acl_tb_mem.c.o 00:03:23.814 [187/740] Generating lib/rte_bbdev_def with a custom command 00:03:24.071 [188/740] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev.c.o 00:03:24.071 [189/740] Compiling C object lib/librte_acl.a.p/acl_rte_acl.c.o 00:03:24.071 [190/740] Generating lib/rte_bbdev_mingw with a custom command 00:03:24.071 [191/740] Linking static target lib/librte_ethdev.a 00:03:24.071 [192/740] Generating lib/rte_bitratestats_def with a custom command 00:03:24.071 [193/740] Generating lib/rte_bitratestats_mingw with a custom command 00:03:24.329 [194/740] Compiling C object lib/librte_bitratestats.a.p/bitratestats_rte_bitrate.c.o 00:03:24.329 [195/740] Linking static target lib/librte_bitratestats.a 00:03:24.329 [196/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf.c.o 00:03:24.329 [197/740] Compiling C object lib/librte_acl.a.p/acl_acl_bld.c.o 00:03:24.587 [198/740] Compiling C object lib/librte_bbdev.a.p/bbdev_rte_bbdev.c.o 00:03:24.587 [199/740] Linking static target lib/librte_bbdev.a 00:03:24.587 [200/740] Generating lib/bitratestats.sym_chk with a custom command (wrapped by meson to capture output) 00:03:24.845 [201/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_dump.c.o 00:03:25.104 [202/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_load.c.o 00:03:25.104 [203/740] Generating lib/bbdev.sym_chk with a custom command (wrapped by meson to capture output) 00:03:25.104 [204/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_exec.c.o 00:03:25.362 [205/740] Compiling C object lib/librte_hash.a.p/hash_rte_cuckoo_hash.c.o 00:03:25.362 [206/740] Linking static target lib/librte_hash.a 00:03:25.362 [207/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_stub.c.o 00:03:25.362 [208/740] Compiling C object lib/librte_acl.a.p/acl_acl_run_sse.c.o 00:03:25.621 [209/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_pkt.c.o 00:03:25.621 [210/740] Generating lib/rte_bpf_def with a custom command 00:03:25.621 [211/740] Generating lib/rte_bpf_mingw with a custom command 00:03:25.878 [212/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_load_elf.c.o 00:03:25.878 [213/740] Generating lib/rte_cfgfile_def with a custom command 00:03:25.878 [214/740] Generating lib/rte_cfgfile_mingw with a custom command 00:03:25.878 [215/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_validate.c.o 00:03:25.878 [216/740] Compiling C object lib/librte_cfgfile.a.p/cfgfile_rte_cfgfile.c.o 00:03:25.878 [217/740] Linking static target lib/librte_cfgfile.a 00:03:25.878 [218/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_convert.c.o 00:03:25.878 [219/740] Generating lib/hash.sym_chk with a custom command (wrapped by meson to capture output) 00:03:25.878 [220/740] Generating lib/rte_compressdev_def with a custom command 00:03:26.137 [221/740] Generating lib/rte_compressdev_mingw with a custom command 00:03:26.137 [222/740] Compiling C object lib/librte_acl.a.p/acl_acl_run_avx2.c.o 00:03:26.137 [223/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_jit_x86.c.o 00:03:26.137 [224/740] Linking static target lib/librte_bpf.a 00:03:26.137 [225/740] Generating lib/cfgfile.sym_chk with a custom command (wrapped by meson to capture output) 00:03:26.395 [226/740] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_compressdev_pmd.c.o 00:03:26.395 [227/740] Generating lib/rte_cryptodev_def with a custom command 00:03:26.395 [228/740] Generating lib/rte_cryptodev_mingw with a custom command 00:03:26.395 [229/740] Compiling C object lib/librte_acl.a.p/acl_acl_run_avx512.c.o 00:03:26.395 [230/740] Generating lib/bpf.sym_chk with a custom command (wrapped by meson to capture output) 00:03:26.395 [231/740] Linking static target lib/librte_acl.a 00:03:26.395 [232/740] Compiling C object lib/librte_cryptodev.a.p/cryptodev_cryptodev_pmd.c.o 00:03:26.395 [233/740] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_compressdev.c.o 00:03:26.395 [234/740] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_comp.c.o 00:03:26.395 [235/740] Linking static target lib/librte_compressdev.a 00:03:26.395 [236/740] Generating lib/rte_distributor_def with a custom command 00:03:26.395 [237/740] Generating lib/rte_distributor_mingw with a custom command 00:03:26.654 [238/740] Compiling C object lib/librte_cryptodev.a.p/cryptodev_cryptodev_trace_points.c.o 00:03:26.654 [239/740] Generating lib/rte_efd_def with a custom command 00:03:26.654 [240/740] Generating lib/acl.sym_chk with a custom command (wrapped by meson to capture output) 00:03:26.654 [241/740] Generating lib/rte_efd_mingw with a custom command 00:03:26.912 [242/740] Compiling C object lib/librte_distributor.a.p/distributor_rte_distributor_match_sse.c.o 00:03:26.912 [243/740] Compiling C object lib/librte_distributor.a.p/distributor_rte_distributor_single.c.o 00:03:27.170 [244/740] Compiling C object lib/librte_eventdev.a.p/eventdev_eventdev_private.c.o 00:03:27.170 [245/740] Compiling C object lib/librte_distributor.a.p/distributor_rte_distributor.c.o 00:03:27.170 [246/740] Linking static target lib/librte_distributor.a 00:03:27.170 [247/740] Compiling C object lib/librte_eventdev.a.p/eventdev_eventdev_trace_points.c.o 00:03:27.170 [248/740] Generating lib/eal.sym_chk with a custom command (wrapped by meson to capture output) 00:03:27.170 [249/740] Generating lib/compressdev.sym_chk with a custom command (wrapped by meson to capture output) 00:03:27.170 [250/740] Linking target lib/librte_eal.so.23.0 00:03:27.442 [251/740] Generating lib/distributor.sym_chk with a custom command (wrapped by meson to capture output) 00:03:27.442 [252/740] Generating symbol file lib/librte_eal.so.23.0.p/librte_eal.so.23.0.symbols 00:03:27.442 [253/740] Linking target lib/librte_ring.so.23.0 00:03:27.442 [254/740] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_ring.c.o 00:03:27.442 [255/740] Generating symbol file lib/librte_ring.so.23.0.p/librte_ring.so.23.0.symbols 00:03:27.733 [256/740] Linking target lib/librte_meter.so.23.0 00:03:27.733 [257/740] Linking target lib/librte_rcu.so.23.0 00:03:27.733 [258/740] Linking target lib/librte_mempool.so.23.0 00:03:27.733 [259/740] Generating symbol file lib/librte_rcu.so.23.0.p/librte_rcu.so.23.0.symbols 00:03:27.733 [260/740] Compiling C object lib/librte_efd.a.p/efd_rte_efd.c.o 00:03:27.733 [261/740] Generating symbol file lib/librte_mempool.so.23.0.p/librte_mempool.so.23.0.symbols 00:03:27.733 [262/740] Linking target lib/librte_pci.so.23.0 00:03:27.733 [263/740] Linking target lib/librte_mbuf.so.23.0 00:03:27.733 [264/740] Generating symbol file lib/librte_meter.so.23.0.p/librte_meter.so.23.0.symbols 00:03:27.733 [265/740] Linking target lib/librte_timer.so.23.0 00:03:27.992 [266/740] Generating symbol file lib/librte_pci.so.23.0.p/librte_pci.so.23.0.symbols 00:03:27.992 [267/740] Generating symbol file lib/librte_mbuf.so.23.0.p/librte_mbuf.so.23.0.symbols 00:03:27.992 [268/740] Linking target lib/librte_acl.so.23.0 00:03:27.992 [269/740] Linking target lib/librte_net.so.23.0 00:03:27.992 [270/740] Generating symbol file lib/librte_timer.so.23.0.p/librte_timer.so.23.0.symbols 00:03:27.992 [271/740] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_crypto_adapter.c.o 00:03:27.992 [272/740] Linking target lib/librte_bbdev.so.23.0 00:03:27.992 [273/740] Linking target lib/librte_cfgfile.so.23.0 00:03:27.992 [274/740] Linking target lib/librte_compressdev.so.23.0 00:03:27.992 [275/740] Linking static target lib/librte_efd.a 00:03:27.992 [276/740] Generating symbol file lib/librte_acl.so.23.0.p/librte_acl.so.23.0.symbols 00:03:27.992 [277/740] Generating symbol file lib/librte_net.so.23.0.p/librte_net.so.23.0.symbols 00:03:27.992 [278/740] Linking target lib/librte_distributor.so.23.0 00:03:27.992 [279/740] Generating lib/rte_eventdev_def with a custom command 00:03:27.992 [280/740] Linking target lib/librte_cmdline.so.23.0 00:03:27.992 [281/740] Generating lib/rte_eventdev_mingw with a custom command 00:03:27.992 [282/740] Linking target lib/librte_hash.so.23.0 00:03:28.251 [283/740] Generating lib/rte_gpudev_def with a custom command 00:03:28.251 [284/740] Generating lib/rte_gpudev_mingw with a custom command 00:03:28.251 [285/740] Compiling C object lib/librte_cryptodev.a.p/cryptodev_rte_cryptodev.c.o 00:03:28.251 [286/740] Linking static target lib/librte_cryptodev.a 00:03:28.251 [287/740] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_eth_tx_adapter.c.o 00:03:28.251 [288/740] Generating symbol file lib/librte_hash.so.23.0.p/librte_hash.so.23.0.symbols 00:03:28.251 [289/740] Generating lib/ethdev.sym_chk with a custom command (wrapped by meson to capture output) 00:03:28.251 [290/740] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_timer_adapter.c.o 00:03:28.251 [291/740] Generating lib/efd.sym_chk with a custom command (wrapped by meson to capture output) 00:03:28.251 [292/740] Linking target lib/librte_ethdev.so.23.0 00:03:28.251 [293/740] Linking target lib/librte_efd.so.23.0 00:03:28.510 [294/740] Generating symbol file lib/librte_ethdev.so.23.0.p/librte_ethdev.so.23.0.symbols 00:03:28.510 [295/740] Linking target lib/librte_metrics.so.23.0 00:03:28.510 [296/740] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_eventdev.c.o 00:03:28.510 [297/740] Linking target lib/librte_bpf.so.23.0 00:03:28.510 [298/740] Generating symbol file lib/librte_metrics.so.23.0.p/librte_metrics.so.23.0.symbols 00:03:28.770 [299/740] Linking target lib/librte_bitratestats.so.23.0 00:03:28.770 [300/740] Generating lib/rte_gro_def with a custom command 00:03:28.770 [301/740] Generating lib/rte_gro_mingw with a custom command 00:03:28.770 [302/740] Generating symbol file lib/librte_bpf.so.23.0.p/librte_bpf.so.23.0.symbols 00:03:28.770 [303/740] Compiling C object lib/librte_gro.a.p/gro_rte_gro.c.o 00:03:28.770 [304/740] Compiling C object lib/librte_gpudev.a.p/gpudev_gpudev.c.o 00:03:28.770 [305/740] Linking static target lib/librte_gpudev.a 00:03:28.770 [306/740] Compiling C object lib/librte_gro.a.p/gro_gro_tcp4.c.o 00:03:28.770 [307/740] Compiling C object lib/librte_gro.a.p/gro_gro_udp4.c.o 00:03:29.028 [308/740] Compiling C object lib/librte_gro.a.p/gro_gro_vxlan_tcp4.c.o 00:03:29.028 [309/740] Compiling C object lib/librte_gro.a.p/gro_gro_vxlan_udp4.c.o 00:03:29.028 [310/740] Linking static target lib/librte_gro.a 00:03:29.287 [311/740] Generating lib/gro.sym_chk with a custom command (wrapped by meson to capture output) 00:03:29.287 [312/740] Compiling C object lib/librte_gso.a.p/gso_gso_tcp4.c.o 00:03:29.287 [313/740] Generating lib/rte_gso_def with a custom command 00:03:29.287 [314/740] Compiling C object lib/librte_gso.a.p/gso_gso_udp4.c.o 00:03:29.287 [315/740] Linking target lib/librte_gro.so.23.0 00:03:29.287 [316/740] Generating lib/rte_gso_mingw with a custom command 00:03:29.287 [317/740] Compiling C object lib/librte_gso.a.p/gso_gso_tunnel_tcp4.c.o 00:03:29.287 [318/740] Compiling C object lib/librte_gso.a.p/gso_gso_tunnel_udp4.c.o 00:03:29.287 [319/740] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_eth_rx_adapter.c.o 00:03:29.287 [320/740] Compiling C object lib/librte_gso.a.p/gso_gso_common.c.o 00:03:29.287 [321/740] Linking static target lib/librte_eventdev.a 00:03:29.287 [322/740] Compiling C object lib/librte_gso.a.p/gso_rte_gso.c.o 00:03:29.287 [323/740] Linking static target lib/librte_gso.a 00:03:29.546 [324/740] Generating lib/gso.sym_chk with a custom command (wrapped by meson to capture output) 00:03:29.546 [325/740] Linking target lib/librte_gso.so.23.0 00:03:29.546 [326/740] Generating lib/gpudev.sym_chk with a custom command (wrapped by meson to capture output) 00:03:29.546 [327/740] Generating lib/rte_ip_frag_def with a custom command 00:03:29.546 [328/740] Linking target lib/librte_gpudev.so.23.0 00:03:29.546 [329/740] Generating lib/rte_ip_frag_mingw with a custom command 00:03:29.546 [330/740] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ipv4_reassembly.c.o 00:03:29.546 [331/740] Generating lib/rte_jobstats_def with a custom command 00:03:29.546 [332/740] Generating lib/rte_jobstats_mingw with a custom command 00:03:29.805 [333/740] Generating lib/rte_latencystats_def with a custom command 00:03:29.805 [334/740] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ipv6_reassembly.c.o 00:03:29.805 [335/740] Generating lib/rte_latencystats_mingw with a custom command 00:03:29.805 [336/740] Compiling C object lib/librte_jobstats.a.p/jobstats_rte_jobstats.c.o 00:03:29.805 [337/740] Linking static target lib/librte_jobstats.a 00:03:29.805 [338/740] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ipv6_fragmentation.c.o 00:03:29.805 [339/740] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ip_frag_common.c.o 00:03:29.805 [340/740] Generating lib/rte_lpm_def with a custom command 00:03:29.805 [341/740] Generating lib/rte_lpm_mingw with a custom command 00:03:29.805 [342/740] Compiling C object lib/librte_ip_frag.a.p/ip_frag_ip_frag_internal.c.o 00:03:30.064 [343/740] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ipv4_fragmentation.c.o 00:03:30.064 [344/740] Linking static target lib/librte_ip_frag.a 00:03:30.064 [345/740] Generating lib/jobstats.sym_chk with a custom command (wrapped by meson to capture output) 00:03:30.064 [346/740] Linking target lib/librte_jobstats.so.23.0 00:03:30.324 [347/740] Generating lib/cryptodev.sym_chk with a custom command (wrapped by meson to capture output) 00:03:30.324 [348/740] Compiling C object lib/librte_latencystats.a.p/latencystats_rte_latencystats.c.o 00:03:30.324 [349/740] Linking static target lib/librte_latencystats.a 00:03:30.324 [350/740] Generating lib/ip_frag.sym_chk with a custom command (wrapped by meson to capture output) 00:03:30.324 [351/740] Linking target lib/librte_cryptodev.so.23.0 00:03:30.324 [352/740] Linking target lib/librte_ip_frag.so.23.0 00:03:30.324 [353/740] Compiling C object lib/librte_lpm.a.p/lpm_rte_lpm.c.o 00:03:30.324 [354/740] Compiling C object lib/librte_member.a.p/member_rte_member.c.o 00:03:30.324 [355/740] Generating symbol file lib/librte_cryptodev.so.23.0.p/librte_cryptodev.so.23.0.symbols 00:03:30.324 [356/740] Compiling C object lib/member/libsketch_avx512_tmp.a.p/rte_member_sketch_avx512.c.o 00:03:30.324 [357/740] Generating lib/rte_member_def with a custom command 00:03:30.324 [358/740] Linking static target lib/member/libsketch_avx512_tmp.a 00:03:30.324 [359/740] Generating symbol file lib/librte_ip_frag.so.23.0.p/librte_ip_frag.so.23.0.symbols 00:03:30.324 [360/740] Generating lib/rte_member_mingw with a custom command 00:03:30.324 [361/740] Generating lib/rte_pcapng_mingw with a custom command 00:03:30.324 [362/740] Generating lib/latencystats.sym_chk with a custom command (wrapped by meson to capture output) 00:03:30.324 [363/740] Generating lib/rte_pcapng_def with a custom command 00:03:30.324 [364/740] Linking target lib/librte_latencystats.so.23.0 00:03:30.583 [365/740] Compiling C object lib/librte_power.a.p/power_guest_channel.c.o 00:03:30.583 [366/740] Compiling C object lib/librte_power.a.p/power_power_common.c.o 00:03:30.583 [367/740] Compiling C object lib/librte_power.a.p/power_power_kvm_vm.c.o 00:03:30.583 [368/740] Compiling C object lib/librte_lpm.a.p/lpm_rte_lpm6.c.o 00:03:30.583 [369/740] Compiling C object lib/librte_power.a.p/power_rte_power.c.o 00:03:30.583 [370/740] Linking static target lib/librte_lpm.a 00:03:30.842 [371/740] Compiling C object lib/librte_member.a.p/member_rte_member_ht.c.o 00:03:30.842 [372/740] Compiling C object lib/librte_power.a.p/power_rte_power_empty_poll.c.o 00:03:30.842 [373/740] Compiling C object lib/librte_power.a.p/power_power_acpi_cpufreq.c.o 00:03:30.842 [374/740] Generating lib/rte_power_def with a custom command 00:03:30.842 [375/740] Compiling C object lib/librte_member.a.p/member_rte_member_vbf.c.o 00:03:30.842 [376/740] Generating lib/rte_power_mingw with a custom command 00:03:31.101 [377/740] Compiling C object lib/librte_power.a.p/power_power_cppc_cpufreq.c.o 00:03:31.101 [378/740] Generating lib/rte_rawdev_def with a custom command 00:03:31.101 [379/740] Generating lib/rte_rawdev_mingw with a custom command 00:03:31.101 [380/740] Generating lib/lpm.sym_chk with a custom command (wrapped by meson to capture output) 00:03:31.101 [381/740] Generating lib/rte_regexdev_def with a custom command 00:03:31.101 [382/740] Linking target lib/librte_lpm.so.23.0 00:03:31.101 [383/740] Compiling C object lib/librte_pcapng.a.p/pcapng_rte_pcapng.c.o 00:03:31.101 [384/740] Generating lib/rte_regexdev_mingw with a custom command 00:03:31.101 [385/740] Linking static target lib/librte_pcapng.a 00:03:31.101 [386/740] Generating lib/eventdev.sym_chk with a custom command (wrapped by meson to capture output) 00:03:31.101 [387/740] Linking target lib/librte_eventdev.so.23.0 00:03:31.101 [388/740] Generating symbol file lib/librte_lpm.so.23.0.p/librte_lpm.so.23.0.symbols 00:03:31.101 [389/740] Generating lib/rte_dmadev_def with a custom command 00:03:31.359 [390/740] Compiling C object lib/librte_power.a.p/power_power_pstate_cpufreq.c.o 00:03:31.359 [391/740] Generating lib/rte_dmadev_mingw with a custom command 00:03:31.359 [392/740] Generating symbol file lib/librte_eventdev.so.23.0.p/librte_eventdev.so.23.0.symbols 00:03:31.359 [393/740] Compiling C object lib/librte_rawdev.a.p/rawdev_rte_rawdev.c.o 00:03:31.359 [394/740] Linking static target lib/librte_rawdev.a 00:03:31.359 [395/740] Generating lib/rte_rib_def with a custom command 00:03:31.359 [396/740] Compiling C object lib/librte_power.a.p/power_rte_power_intel_uncore.c.o 00:03:31.359 [397/740] Generating lib/rte_rib_mingw with a custom command 00:03:31.359 [398/740] Generating lib/rte_reorder_def with a custom command 00:03:31.359 [399/740] Generating lib/rte_reorder_mingw with a custom command 00:03:31.359 [400/740] Generating lib/pcapng.sym_chk with a custom command (wrapped by meson to capture output) 00:03:31.359 [401/740] Compiling C object lib/librte_power.a.p/power_rte_power_pmd_mgmt.c.o 00:03:31.359 [402/740] Linking static target lib/librte_power.a 00:03:31.359 [403/740] Linking target lib/librte_pcapng.so.23.0 00:03:31.359 [404/740] Compiling C object lib/librte_dmadev.a.p/dmadev_rte_dmadev.c.o 00:03:31.359 [405/740] Linking static target lib/librte_dmadev.a 00:03:31.617 [406/740] Compiling C object lib/librte_regexdev.a.p/regexdev_rte_regexdev.c.o 00:03:31.617 [407/740] Linking static target lib/librte_regexdev.a 00:03:31.617 [408/740] Generating symbol file lib/librte_pcapng.so.23.0.p/librte_pcapng.so.23.0.symbols 00:03:31.617 [409/740] Compiling C object lib/librte_sched.a.p/sched_rte_red.c.o 00:03:31.617 [410/740] Generating lib/rawdev.sym_chk with a custom command (wrapped by meson to capture output) 00:03:31.875 [411/740] Linking target lib/librte_rawdev.so.23.0 00:03:31.875 [412/740] Compiling C object lib/librte_sched.a.p/sched_rte_approx.c.o 00:03:31.875 [413/740] Compiling C object lib/librte_rib.a.p/rib_rte_rib.c.o 00:03:31.875 [414/740] Generating lib/rte_sched_def with a custom command 00:03:31.875 [415/740] Generating lib/rte_sched_mingw with a custom command 00:03:31.875 [416/740] Compiling C object lib/librte_sched.a.p/sched_rte_pie.c.o 00:03:31.875 [417/740] Generating lib/rte_security_def with a custom command 00:03:31.875 [418/740] Generating lib/rte_security_mingw with a custom command 00:03:31.875 [419/740] Compiling C object lib/librte_stack.a.p/stack_rte_stack_std.c.o 00:03:31.875 [420/740] Compiling C object lib/librte_reorder.a.p/reorder_rte_reorder.c.o 00:03:31.875 [421/740] Linking static target lib/librte_reorder.a 00:03:31.875 [422/740] Generating lib/dmadev.sym_chk with a custom command (wrapped by meson to capture output) 00:03:31.875 [423/740] Compiling C object lib/librte_stack.a.p/stack_rte_stack.c.o 00:03:31.875 [424/740] Compiling C object lib/librte_member.a.p/member_rte_member_sketch.c.o 00:03:31.875 [425/740] Linking target lib/librte_dmadev.so.23.0 00:03:31.875 [426/740] Linking static target lib/librte_member.a 00:03:32.133 [427/740] Compiling C object lib/librte_stack.a.p/stack_rte_stack_lf.c.o 00:03:32.133 [428/740] Generating lib/rte_stack_def with a custom command 00:03:32.133 [429/740] Linking static target lib/librte_stack.a 00:03:32.133 [430/740] Generating lib/rte_stack_mingw with a custom command 00:03:32.133 [431/740] Generating symbol file lib/librte_dmadev.so.23.0.p/librte_dmadev.so.23.0.symbols 00:03:32.133 [432/740] Compiling C object lib/librte_rib.a.p/rib_rte_rib6.c.o 00:03:32.133 [433/740] Linking static target lib/librte_rib.a 00:03:32.133 [434/740] Compiling C object lib/librte_vhost.a.p/vhost_fd_man.c.o 00:03:32.133 [435/740] Generating lib/stack.sym_chk with a custom command (wrapped by meson to capture output) 00:03:32.133 [436/740] Generating lib/reorder.sym_chk with a custom command (wrapped by meson to capture output) 00:03:32.133 [437/740] Generating lib/regexdev.sym_chk with a custom command (wrapped by meson to capture output) 00:03:32.133 [438/740] Linking target lib/librte_stack.so.23.0 00:03:32.133 [439/740] Linking target lib/librte_reorder.so.23.0 00:03:32.133 [440/740] Linking target lib/librte_regexdev.so.23.0 00:03:32.392 [441/740] Generating lib/power.sym_chk with a custom command (wrapped by meson to capture output) 00:03:32.392 [442/740] Generating lib/member.sym_chk with a custom command (wrapped by meson to capture output) 00:03:32.392 [443/740] Linking target lib/librte_power.so.23.0 00:03:32.392 [444/740] Linking target lib/librte_member.so.23.0 00:03:32.392 [445/740] Compiling C object lib/librte_security.a.p/security_rte_security.c.o 00:03:32.392 [446/740] Linking static target lib/librte_security.a 00:03:32.392 [447/740] Generating lib/rib.sym_chk with a custom command (wrapped by meson to capture output) 00:03:32.392 [448/740] Linking target lib/librte_rib.so.23.0 00:03:32.650 [449/740] Compiling C object lib/librte_vhost.a.p/vhost_iotlb.c.o 00:03:32.650 [450/740] Generating lib/rte_vhost_def with a custom command 00:03:32.650 [451/740] Generating symbol file lib/librte_rib.so.23.0.p/librte_rib.so.23.0.symbols 00:03:32.650 [452/740] Generating lib/rte_vhost_mingw with a custom command 00:03:32.650 [453/740] Compiling C object lib/librte_vhost.a.p/vhost_vdpa.c.o 00:03:32.908 [454/740] Compiling C object lib/librte_vhost.a.p/vhost_socket.c.o 00:03:32.908 [455/740] Compiling C object lib/librte_sched.a.p/sched_rte_sched.c.o 00:03:32.908 [456/740] Generating lib/security.sym_chk with a custom command (wrapped by meson to capture output) 00:03:32.908 [457/740] Linking static target lib/librte_sched.a 00:03:32.908 [458/740] Linking target lib/librte_security.so.23.0 00:03:32.908 [459/740] Generating symbol file lib/librte_security.so.23.0.p/librte_security.so.23.0.symbols 00:03:33.167 [460/740] Generating lib/sched.sym_chk with a custom command (wrapped by meson to capture output) 00:03:33.167 [461/740] Compiling C object lib/librte_ipsec.a.p/ipsec_ses.c.o 00:03:33.167 [462/740] Linking target lib/librte_sched.so.23.0 00:03:33.167 [463/740] Compiling C object lib/librte_ipsec.a.p/ipsec_sa.c.o 00:03:33.425 [464/740] Generating lib/rte_ipsec_def with a custom command 00:03:33.425 [465/740] Generating lib/rte_ipsec_mingw with a custom command 00:03:33.425 [466/740] Generating symbol file lib/librte_sched.so.23.0.p/librte_sched.so.23.0.symbols 00:03:33.425 [467/740] Compiling C object lib/librte_vhost.a.p/vhost_vhost.c.o 00:03:33.425 [468/740] Compiling C object lib/librte_fib.a.p/fib_rte_fib.c.o 00:03:33.425 [469/740] Compiling C object lib/librte_vhost.a.p/vhost_vhost_user.c.o 00:03:33.683 [470/740] Compiling C object lib/librte_ipsec.a.p/ipsec_ipsec_telemetry.c.o 00:03:33.683 [471/740] Compiling C object lib/librte_ipsec.a.p/ipsec_ipsec_sad.c.o 00:03:33.683 [472/740] Generating lib/rte_fib_def with a custom command 00:03:33.683 [473/740] Generating lib/rte_fib_mingw with a custom command 00:03:33.683 [474/740] Compiling C object lib/librte_fib.a.p/fib_rte_fib6.c.o 00:03:33.942 [475/740] Compiling C object lib/librte_fib.a.p/fib_trie_avx512.c.o 00:03:33.942 [476/740] Compiling C object lib/librte_fib.a.p/fib_dir24_8_avx512.c.o 00:03:34.201 [477/740] Compiling C object lib/librte_fib.a.p/fib_trie.c.o 00:03:34.201 [478/740] Compiling C object lib/librte_ipsec.a.p/ipsec_esp_outb.c.o 00:03:34.201 [479/740] Compiling C object lib/librte_ipsec.a.p/ipsec_esp_inb.c.o 00:03:34.460 [480/740] Linking static target lib/librte_ipsec.a 00:03:34.460 [481/740] Compiling C object lib/librte_port.a.p/port_rte_port_ethdev.c.o 00:03:34.460 [482/740] Compiling C object lib/librte_fib.a.p/fib_dir24_8.c.o 00:03:34.460 [483/740] Linking static target lib/librte_fib.a 00:03:34.460 [484/740] Compiling C object lib/librte_port.a.p/port_rte_port_fd.c.o 00:03:34.718 [485/740] Compiling C object lib/librte_port.a.p/port_rte_port_frag.c.o 00:03:34.718 [486/740] Compiling C object lib/librte_port.a.p/port_rte_port_sched.c.o 00:03:34.718 [487/740] Generating lib/ipsec.sym_chk with a custom command (wrapped by meson to capture output) 00:03:34.718 [488/740] Compiling C object lib/librte_port.a.p/port_rte_port_ras.c.o 00:03:34.718 [489/740] Linking target lib/librte_ipsec.so.23.0 00:03:34.718 [490/740] Generating lib/fib.sym_chk with a custom command (wrapped by meson to capture output) 00:03:34.977 [491/740] Linking target lib/librte_fib.so.23.0 00:03:35.236 [492/740] Compiling C object lib/librte_port.a.p/port_rte_port_source_sink.c.o 00:03:35.236 [493/740] Generating lib/rte_port_def with a custom command 00:03:35.236 [494/740] Generating lib/rte_port_mingw with a custom command 00:03:35.236 [495/740] Compiling C object lib/librte_port.a.p/port_rte_port_sym_crypto.c.o 00:03:35.236 [496/740] Compiling C object lib/librte_port.a.p/port_rte_swx_port_ethdev.c.o 00:03:35.236 [497/740] Generating lib/rte_pdump_def with a custom command 00:03:35.236 [498/740] Generating lib/rte_pdump_mingw with a custom command 00:03:35.236 [499/740] Compiling C object lib/librte_port.a.p/port_rte_swx_port_fd.c.o 00:03:35.498 [500/740] Compiling C object lib/librte_port.a.p/port_rte_port_eventdev.c.o 00:03:35.498 [501/740] Compiling C object lib/librte_table.a.p/table_rte_swx_keycmp.c.o 00:03:35.498 [502/740] Compiling C object lib/librte_port.a.p/port_rte_swx_port_source_sink.c.o 00:03:35.498 [503/740] Compiling C object lib/librte_table.a.p/table_rte_swx_table_em.c.o 00:03:35.498 [504/740] Compiling C object lib/librte_table.a.p/table_rte_swx_table_learner.c.o 00:03:35.756 [505/740] Compiling C object lib/librte_port.a.p/port_rte_swx_port_ring.c.o 00:03:35.756 [506/740] Compiling C object lib/librte_table.a.p/table_rte_swx_table_selector.c.o 00:03:35.756 [507/740] Compiling C object lib/librte_table.a.p/table_rte_swx_table_wm.c.o 00:03:36.014 [508/740] Compiling C object lib/librte_port.a.p/port_rte_port_ring.c.o 00:03:36.014 [509/740] Compiling C object lib/librte_table.a.p/table_rte_table_array.c.o 00:03:36.014 [510/740] Linking static target lib/librte_port.a 00:03:36.014 [511/740] Compiling C object lib/librte_table.a.p/table_rte_table_hash_cuckoo.c.o 00:03:36.014 [512/740] Compiling C object lib/librte_table.a.p/table_rte_table_acl.c.o 00:03:36.014 [513/740] Compiling C object lib/librte_pdump.a.p/pdump_rte_pdump.c.o 00:03:36.014 [514/740] Linking static target lib/librte_pdump.a 00:03:36.273 [515/740] Generating lib/pdump.sym_chk with a custom command (wrapped by meson to capture output) 00:03:36.273 [516/740] Generating lib/port.sym_chk with a custom command (wrapped by meson to capture output) 00:03:36.273 [517/740] Linking target lib/librte_pdump.so.23.0 00:03:36.273 [518/740] Linking target lib/librte_port.so.23.0 00:03:36.533 [519/740] Compiling C object lib/librte_table.a.p/table_rte_table_lpm.c.o 00:03:36.533 [520/740] Compiling C object lib/librte_table.a.p/table_rte_table_hash_ext.c.o 00:03:36.533 [521/740] Generating symbol file lib/librte_port.so.23.0.p/librte_port.so.23.0.symbols 00:03:36.533 [522/740] Generating lib/rte_table_def with a custom command 00:03:36.533 [523/740] Generating lib/rte_table_mingw with a custom command 00:03:36.533 [524/740] Compiling C object lib/librte_table.a.p/table_rte_table_hash_key8.c.o 00:03:36.792 [525/740] Compiling C object lib/librte_table.a.p/table_rte_table_hash_key16.c.o 00:03:36.792 [526/740] Compiling C object lib/librte_table.a.p/table_rte_table_lpm_ipv6.c.o 00:03:36.792 [527/740] Compiling C object lib/librte_table.a.p/table_rte_table_stub.c.o 00:03:36.792 [528/740] Compiling C object lib/librte_table.a.p/table_rte_table_hash_key32.c.o 00:03:36.792 [529/740] Generating lib/rte_pipeline_def with a custom command 00:03:36.792 [530/740] Generating lib/rte_pipeline_mingw with a custom command 00:03:36.792 [531/740] Compiling C object lib/librte_table.a.p/table_rte_table_hash_lru.c.o 00:03:37.051 [532/740] Linking static target lib/librte_table.a 00:03:37.051 [533/740] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_port_in_action.c.o 00:03:37.311 [534/740] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_pipeline.c.o 00:03:37.311 [535/740] Compiling C object lib/librte_graph.a.p/graph_node.c.o 00:03:37.311 [536/740] Generating lib/table.sym_chk with a custom command (wrapped by meson to capture output) 00:03:37.311 [537/740] Compiling C object lib/librte_graph.a.p/graph_graph.c.o 00:03:37.570 [538/740] Linking target lib/librte_table.so.23.0 00:03:37.570 [539/740] Generating symbol file lib/librte_table.so.23.0.p/librte_table.so.23.0.symbols 00:03:37.570 [540/740] Compiling C object lib/librte_graph.a.p/graph_graph_ops.c.o 00:03:37.570 [541/740] Generating lib/rte_graph_def with a custom command 00:03:37.570 [542/740] Generating lib/rte_graph_mingw with a custom command 00:03:37.570 [543/740] Compiling C object lib/librte_vhost.a.p/vhost_vhost_crypto.c.o 00:03:37.570 [544/740] Compiling C object lib/librte_graph.a.p/graph_graph_debug.c.o 00:03:37.830 [545/740] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_swx_ctl.c.o 00:03:37.830 [546/740] Compiling C object lib/librte_graph.a.p/graph_graph_stats.c.o 00:03:37.830 [547/740] Compiling C object lib/librte_graph.a.p/graph_graph_populate.c.o 00:03:37.830 [548/740] Linking static target lib/librte_graph.a 00:03:38.089 [549/740] Compiling C object lib/librte_node.a.p/node_ethdev_ctrl.c.o 00:03:38.089 [550/740] Compiling C object lib/librte_node.a.p/node_ethdev_tx.c.o 00:03:38.089 [551/740] Compiling C object lib/librte_node.a.p/node_null.c.o 00:03:38.089 [552/740] Compiling C object lib/librte_node.a.p/node_ethdev_rx.c.o 00:03:38.348 [553/740] Compiling C object lib/librte_node.a.p/node_log.c.o 00:03:38.348 [554/740] Generating lib/rte_node_def with a custom command 00:03:38.348 [555/740] Generating lib/rte_node_mingw with a custom command 00:03:38.607 [556/740] Compiling C object lib/librte_node.a.p/node_pkt_drop.c.o 00:03:38.607 [557/740] Generating lib/graph.sym_chk with a custom command (wrapped by meson to capture output) 00:03:38.607 [558/740] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_swx_pipeline_spec.c.o 00:03:38.607 [559/740] Linking target lib/librte_graph.so.23.0 00:03:38.607 [560/740] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_params.c.o 00:03:38.607 [561/740] Generating symbol file lib/librte_graph.so.23.0.p/librte_graph.so.23.0.symbols 00:03:38.607 [562/740] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_common.c.o 00:03:38.607 [563/740] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_common_uio.c.o 00:03:38.607 [564/740] Compiling C object lib/librte_node.a.p/node_ip4_lookup.c.o 00:03:38.607 [565/740] Generating drivers/rte_bus_pci_def with a custom command 00:03:38.607 [566/740] Generating drivers/rte_bus_pci_mingw with a custom command 00:03:38.866 [567/740] Compiling C object drivers/libtmp_rte_bus_vdev.a.p/bus_vdev_vdev_params.c.o 00:03:38.866 [568/740] Compiling C object lib/librte_node.a.p/node_pkt_cls.c.o 00:03:38.866 [569/740] Generating drivers/rte_bus_vdev_def with a custom command 00:03:38.866 [570/740] Generating drivers/rte_bus_vdev_mingw with a custom command 00:03:38.866 [571/740] Generating drivers/rte_mempool_ring_def with a custom command 00:03:38.866 [572/740] Generating drivers/rte_mempool_ring_mingw with a custom command 00:03:38.866 [573/740] Compiling C object lib/librte_node.a.p/node_ip4_rewrite.c.o 00:03:38.866 [574/740] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci_uio.c.o 00:03:38.866 [575/740] Linking static target lib/librte_node.a 00:03:38.866 [576/740] Compiling C object drivers/libtmp_rte_bus_vdev.a.p/bus_vdev_vdev.c.o 00:03:38.866 [577/740] Linking static target drivers/libtmp_rte_bus_vdev.a 00:03:38.866 [578/740] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci_vfio.c.o 00:03:39.124 [579/740] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci.c.o 00:03:39.124 [580/740] Linking static target drivers/libtmp_rte_bus_pci.a 00:03:39.124 [581/740] Generating lib/node.sym_chk with a custom command (wrapped by meson to capture output) 00:03:39.124 [582/740] Linking target lib/librte_node.so.23.0 00:03:39.124 [583/740] Generating drivers/rte_bus_vdev.pmd.c with a custom command 00:03:39.124 [584/740] Compiling C object drivers/librte_bus_vdev.a.p/meson-generated_.._rte_bus_vdev.pmd.c.o 00:03:39.124 [585/740] Linking static target drivers/librte_bus_vdev.a 00:03:39.124 [586/740] Generating drivers/rte_bus_pci.pmd.c with a custom command 00:03:39.382 [587/740] Compiling C object drivers/librte_bus_pci.a.p/meson-generated_.._rte_bus_pci.pmd.c.o 00:03:39.382 [588/740] Linking static target drivers/librte_bus_pci.a 00:03:39.382 [589/740] Generating drivers/rte_bus_vdev.sym_chk with a custom command (wrapped by meson to capture output) 00:03:39.382 [590/740] Compiling C object drivers/librte_bus_pci.so.23.0.p/meson-generated_.._rte_bus_pci.pmd.c.o 00:03:39.382 [591/740] Compiling C object drivers/librte_bus_vdev.so.23.0.p/meson-generated_.._rte_bus_vdev.pmd.c.o 00:03:39.382 [592/740] Linking target drivers/librte_bus_vdev.so.23.0 00:03:39.382 [593/740] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_diag.c.o 00:03:39.641 [594/740] Generating symbol file drivers/librte_bus_vdev.so.23.0.p/librte_bus_vdev.so.23.0.symbols 00:03:39.641 [595/740] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_adminq.c.o 00:03:39.641 [596/740] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_dcb.c.o 00:03:39.641 [597/740] Generating drivers/rte_bus_pci.sym_chk with a custom command (wrapped by meson to capture output) 00:03:39.641 [598/740] Linking target drivers/librte_bus_pci.so.23.0 00:03:39.641 [599/740] Compiling C object drivers/libtmp_rte_mempool_ring.a.p/mempool_ring_rte_mempool_ring.c.o 00:03:39.641 [600/740] Linking static target drivers/libtmp_rte_mempool_ring.a 00:03:39.899 [601/740] Generating symbol file drivers/librte_bus_pci.so.23.0.p/librte_bus_pci.so.23.0.symbols 00:03:39.899 [602/740] Generating drivers/rte_mempool_ring.pmd.c with a custom command 00:03:39.899 [603/740] Compiling C object drivers/librte_mempool_ring.a.p/meson-generated_.._rte_mempool_ring.pmd.c.o 00:03:39.899 [604/740] Linking static target drivers/librte_mempool_ring.a 00:03:39.899 [605/740] Compiling C object drivers/librte_mempool_ring.so.23.0.p/meson-generated_.._rte_mempool_ring.pmd.c.o 00:03:39.899 [606/740] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_hmc.c.o 00:03:39.899 [607/740] Linking target drivers/librte_mempool_ring.so.23.0 00:03:40.157 [608/740] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_lan_hmc.c.o 00:03:40.416 [609/740] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_nvm.c.o 00:03:40.674 [610/740] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_common.c.o 00:03:40.674 [611/740] Linking static target drivers/net/i40e/base/libi40e_base.a 00:03:40.932 [612/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_pf.c.o 00:03:41.191 [613/740] Compiling C object drivers/net/i40e/libi40e_avx512_lib.a.p/i40e_rxtx_vec_avx512.c.o 00:03:41.191 [614/740] Linking static target drivers/net/i40e/libi40e_avx512_lib.a 00:03:41.449 [615/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_tm.c.o 00:03:41.706 [616/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_fdir.c.o 00:03:41.706 [617/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_hash.c.o 00:03:41.706 [618/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_flow.c.o 00:03:41.706 [619/740] Generating drivers/rte_net_i40e_def with a custom command 00:03:41.706 [620/740] Generating drivers/rte_net_i40e_mingw with a custom command 00:03:41.706 [621/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_vf_representor.c.o 00:03:42.273 [622/740] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_swx_pipeline.c.o 00:03:42.273 [623/740] Compiling C object app/dpdk-dumpcap.p/dumpcap_main.c.o 00:03:42.666 [624/740] Compiling C object app/dpdk-pdump.p/pdump_main.c.o 00:03:42.666 [625/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_rte_pmd_i40e.c.o 00:03:42.947 [626/740] Compiling C object app/dpdk-test-bbdev.p/test-bbdev_main.c.o 00:03:42.947 [627/740] Compiling C object app/dpdk-test-acl.p/test-acl_main.c.o 00:03:42.947 [628/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_rxtx_vec_sse.c.o 00:03:43.206 [629/740] Compiling C object app/dpdk-test-cmdline.p/test-cmdline_commands.c.o 00:03:43.206 [630/740] Compiling C object app/dpdk-test-cmdline.p/test-cmdline_cmdline_test.c.o 00:03:43.206 [631/740] Compiling C object app/dpdk-proc-info.p/proc-info_main.c.o 00:03:43.206 [632/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_rxtx.c.o 00:03:43.465 [633/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_rxtx_vec_avx2.c.o 00:03:43.723 [634/740] Compiling C object app/dpdk-test-bbdev.p/test-bbdev_test_bbdev.c.o 00:03:43.723 [635/740] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_options_parse.c.o 00:03:43.723 [636/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_ethdev.c.o 00:03:43.723 [637/740] Linking static target drivers/libtmp_rte_net_i40e.a 00:03:43.982 [638/740] Compiling C object app/dpdk-test-bbdev.p/test-bbdev_test_bbdev_vector.c.o 00:03:43.982 [639/740] Generating drivers/rte_net_i40e.pmd.c with a custom command 00:03:43.982 [640/740] Compiling C object drivers/librte_net_i40e.a.p/meson-generated_.._rte_net_i40e.pmd.c.o 00:03:44.241 [641/740] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_main.c.o 00:03:44.241 [642/740] Linking static target drivers/librte_net_i40e.a 00:03:44.241 [643/740] Compiling C object drivers/librte_net_i40e.so.23.0.p/meson-generated_.._rte_net_i40e.pmd.c.o 00:03:44.241 [644/740] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_test_throughput.c.o 00:03:44.241 [645/740] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_test_common.c.o 00:03:44.241 [646/740] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_test_cyclecount.c.o 00:03:44.500 [647/740] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_test_verify.c.o 00:03:44.500 [648/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_common.c.o 00:03:44.758 [649/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_ops.c.o 00:03:44.758 [650/740] Generating drivers/rte_net_i40e.sym_chk with a custom command (wrapped by meson to capture output) 00:03:44.758 [651/740] Linking target drivers/librte_net_i40e.so.23.0 00:03:44.758 [652/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_options_parsing.c.o 00:03:45.018 [653/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_latency.c.o 00:03:45.018 [654/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_vectors.c.o 00:03:45.018 [655/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_pmd_cyclecount.c.o 00:03:45.018 [656/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_throughput.c.o 00:03:45.018 [657/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_evt_test.c.o 00:03:45.018 [658/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_vector_parsing.c.o 00:03:45.276 [659/740] Compiling C object lib/librte_vhost.a.p/vhost_virtio_net.c.o 00:03:45.276 [660/740] Linking static target lib/librte_vhost.a 00:03:45.276 [661/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_parser.c.o 00:03:45.276 [662/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_main.c.o 00:03:45.535 [663/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_evt_main.c.o 00:03:45.535 [664/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_verify.c.o 00:03:45.535 [665/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_evt_options.c.o 00:03:45.793 [666/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_order_common.c.o 00:03:45.793 [667/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_order_atq.c.o 00:03:45.793 [668/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_order_queue.c.o 00:03:46.361 [669/740] Generating lib/vhost.sym_chk with a custom command (wrapped by meson to capture output) 00:03:46.361 [670/740] Linking target lib/librte_vhost.so.23.0 00:03:46.361 [671/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_perf_atq.c.o 00:03:46.361 [672/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_perf_queue.c.o 00:03:46.620 [673/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_pipeline_atq.c.o 00:03:46.620 [674/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_pipeline_common.c.o 00:03:46.620 [675/740] Compiling C object app/dpdk-test-flow-perf.p/test-flow-perf_flow_gen.c.o 00:03:46.880 [676/740] Compiling C object app/dpdk-test-fib.p/test-fib_main.c.o 00:03:46.880 [677/740] Compiling C object app/dpdk-test-flow-perf.p/test-flow-perf_items_gen.c.o 00:03:46.880 [678/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_pipeline_queue.c.o 00:03:46.880 [679/740] Compiling C object app/dpdk-test-flow-perf.p/test-flow-perf_actions_gen.c.o 00:03:47.140 [680/740] Compiling C object app/dpdk-test-gpudev.p/test-gpudev_main.c.o 00:03:47.140 [681/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_config.c.o 00:03:47.140 [682/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_perf_common.c.o 00:03:47.140 [683/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_main.c.o 00:03:47.140 [684/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_init.c.o 00:03:47.140 [685/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_acl.c.o 00:03:47.399 [686/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_hash.c.o 00:03:47.399 [687/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_lpm_ipv6.c.o 00:03:47.399 [688/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_lpm.c.o 00:03:47.659 [689/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_stub.c.o 00:03:47.659 [690/740] Compiling C object app/dpdk-test-flow-perf.p/test-flow-perf_main.c.o 00:03:47.918 [691/740] Compiling C object app/dpdk-test-bbdev.p/test-bbdev_test_bbdev_perf.c.o 00:03:47.918 [692/740] Compiling C object app/dpdk-testpmd.p/test-pmd_5tswap.c.o 00:03:47.918 [693/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_runtime.c.o 00:03:47.918 [694/740] Compiling C object app/dpdk-testpmd.p/test-pmd_cmd_flex_item.c.o 00:03:48.177 [695/740] Compiling C object app/dpdk-testpmd.p/test-pmd_cmdline_mtr.c.o 00:03:48.177 [696/740] Compiling C object app/dpdk-testpmd.p/test-pmd_cmdline_tm.c.o 00:03:48.436 [697/740] Compiling C object app/dpdk-testpmd.p/test-pmd_ieee1588fwd.c.o 00:03:48.436 [698/740] Compiling C object app/dpdk-testpmd.p/test-pmd_icmpecho.c.o 00:03:48.436 [699/740] Compiling C object app/dpdk-testpmd.p/test-pmd_flowgen.c.o 00:03:48.696 [700/740] Compiling C object app/dpdk-testpmd.p/test-pmd_iofwd.c.o 00:03:48.696 [701/740] Compiling C object app/dpdk-testpmd.p/test-pmd_macfwd.c.o 00:03:49.265 [702/740] Compiling C object app/dpdk-testpmd.p/test-pmd_macswap.c.o 00:03:49.265 [703/740] Compiling C object app/dpdk-testpmd.p/test-pmd_csumonly.c.o 00:03:49.265 [704/740] Compiling C object app/dpdk-testpmd.p/test-pmd_cmdline.c.o 00:03:49.265 [705/740] Compiling C object app/dpdk-testpmd.p/test-pmd_shared_rxq_fwd.c.o 00:03:49.265 [706/740] Compiling C object app/dpdk-testpmd.p/test-pmd_rxonly.c.o 00:03:49.524 [707/740] Compiling C object app/dpdk-testpmd.p/test-pmd_parameters.c.o 00:03:49.783 [708/740] Compiling C object app/dpdk-testpmd.p/test-pmd_bpf_cmd.c.o 00:03:50.042 [709/740] Compiling C object app/dpdk-testpmd.p/test-pmd_util.c.o 00:03:50.042 [710/740] Compiling C object app/dpdk-testpmd.p/test-pmd_config.c.o 00:03:50.042 [711/740] Compiling C object app/dpdk-testpmd.p/.._drivers_net_i40e_i40e_testpmd.c.o 00:03:50.302 [712/740] Compiling C object app/dpdk-test-regex.p/test-regex_main.c.o 00:03:50.302 [713/740] Compiling C object app/dpdk-testpmd.p/test-pmd_txonly.c.o 00:03:50.302 [714/740] Compiling C object app/dpdk-test-sad.p/test-sad_main.c.o 00:03:50.302 [715/740] Compiling C object app/dpdk-testpmd.p/test-pmd_cmdline_flow.c.o 00:03:50.561 [716/740] Compiling C object app/dpdk-testpmd.p/test-pmd_noisy_vnf.c.o 00:03:50.561 [717/740] Compiling C object app/dpdk-test-security-perf.p/test-security-perf_test_security_perf.c.o 00:03:50.821 [718/740] Compiling C object app/dpdk-testpmd.p/test-pmd_testpmd.c.o 00:03:51.081 [719/740] Compiling C object app/dpdk-test-security-perf.p/test_test_cryptodev_security_ipsec.c.o 00:03:53.617 [720/740] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_table_action.c.o 00:03:53.617 [721/740] Linking static target lib/librte_pipeline.a 00:03:53.876 [722/740] Linking target app/dpdk-pdump 00:03:53.876 [723/740] Linking target app/dpdk-test-compress-perf 00:03:53.876 [724/740] Linking target app/dpdk-test-crypto-perf 00:03:53.876 [725/740] Linking target app/dpdk-test-eventdev 00:03:53.876 [726/740] Linking target app/dpdk-proc-info 00:03:53.876 [727/740] Linking target app/dpdk-test-cmdline 00:03:53.876 [728/740] Linking target app/dpdk-test-bbdev 00:03:53.876 [729/740] Linking target app/dpdk-test-acl 00:03:53.876 [730/740] Linking target app/dpdk-dumpcap 00:03:54.135 [731/740] Linking target app/dpdk-test-fib 00:03:54.135 [732/740] Linking target app/dpdk-test-flow-perf 00:03:54.135 [733/740] Linking target app/dpdk-test-pipeline 00:03:54.135 [734/740] Linking target app/dpdk-test-regex 00:03:54.135 [735/740] Linking target app/dpdk-test-gpudev 00:03:54.135 [736/740] Linking target app/dpdk-test-sad 00:03:54.135 [737/740] Linking target app/dpdk-test-security-perf 00:03:54.135 [738/740] Linking target app/dpdk-testpmd 00:03:58.329 [739/740] Generating lib/pipeline.sym_chk with a custom command (wrapped by meson to capture output) 00:03:58.588 [740/740] Linking target lib/librte_pipeline.so.23.0 00:03:58.588 15:09:48 build_native_dpdk -- common/autobuild_common.sh@194 -- $ uname -s 00:03:58.588 15:09:48 build_native_dpdk -- common/autobuild_common.sh@194 -- $ [[ Linux == \F\r\e\e\B\S\D ]] 00:03:58.588 15:09:48 build_native_dpdk -- common/autobuild_common.sh@207 -- $ ninja -C /home/vagrant/spdk_repo/dpdk/build-tmp -j10 install 00:03:58.588 ninja: Entering directory `/home/vagrant/spdk_repo/dpdk/build-tmp' 00:03:58.588 [0/1] Installing files. 00:03:58.849 Installing subdir /home/vagrant/spdk_repo/dpdk/examples to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/bbdev_app/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bbdev_app 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/bbdev_app/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bbdev_app 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/bond/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bond 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/bond/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bond 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/bond/main.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bond 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/README to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/dummy.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/t1.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/t2.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/t3.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/commands.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/commands.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/parse_obj_list.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/parse_obj_list.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/common/pkt_group.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/common 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/common/altivec/port_group.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/common/altivec 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/common/neon/port_group.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/common/neon 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/common/sse/port_group.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/common/sse 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/distributor/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/distributor 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/distributor/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/distributor 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/dma/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/dma 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/dma/dmafwd.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/dma 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/ethtool-app/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/ethtool-app 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/ethtool-app/ethapp.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/ethtool-app 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/ethtool-app/ethapp.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/ethtool-app 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/ethtool-app/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/ethtool-app 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/lib/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/lib 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/lib/rte_ethtool.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/lib 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/lib/rte_ethtool.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/lib 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/pipeline_common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/pipeline_worker_generic.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/pipeline_worker_tx.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_dev_self_test.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_dev_self_test.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_aes.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_ccm.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_cmac.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_ecdsa.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_gcm.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:58.849 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_hmac.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_rsa.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_sha.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_tdes.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_xts.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_classify/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_classify 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_classify/flow_classify.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_classify 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_classify/ipv4_rules_file.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_classify 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_filtering/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_filtering 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_filtering/flow_blocks.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_filtering 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_filtering/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_filtering 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/helloworld/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/helloworld 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/helloworld/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/helloworld 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_fragmentation/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_fragmentation 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_fragmentation/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_fragmentation 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/action.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/action.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/cli.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/cli.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/conn.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/conn.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/cryptodev.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/cryptodev.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/kni.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/kni.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/link.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/link.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/mempool.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/mempool.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/parser.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/parser.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/pipeline.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/pipeline.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/swq.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/swq.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/tap.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/tap.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/thread.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/thread.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/tmgr.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/tmgr.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/firewall.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/flow.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/flow_crypto.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/kni.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/l2fwd.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/route.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/route_ecmp.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/rss.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/tap.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_reassembly/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_reassembly 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_reassembly/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_reassembly 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ep0.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ep1.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/esp.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/esp.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/event_helper.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/event_helper.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/flow.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/flow.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipip.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec-secgw.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec-secgw.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_lpm_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_process.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_worker.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_worker.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/parser.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/parser.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/rt.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sa.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sad.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sad.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sp4.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sp6.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/bypass_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/common_defs_secgw.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:58.850 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/data_rxtx.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/linux_test.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/load_env.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/pkttest.py to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/pkttest.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/run_test.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_3descbc_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_3descbc_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aescbc_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aescbc_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aesctr_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aesctr_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aesgcm_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aesgcm_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_ipv6opts.py to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_3descbc_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_3descbc_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aescbc_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aescbc_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aesctr_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aesctr_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aesgcm_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aesgcm_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_null_header_reconstruct.py to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/ipv4_multicast/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipv4_multicast 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/ipv4_multicast/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipv4_multicast 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-cat/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-cat 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-cat/cat.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-cat 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-cat/cat.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-cat 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-cat/l2fwd-cat.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-cat 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-crypto/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-crypto 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-crypto/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-crypto 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_common.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_event.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_event.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_event_generic.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_event_internal_port.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_poll.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_poll.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-jobstats/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-jobstats 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-jobstats/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-jobstats 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/shm.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/shm.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/ka-agent/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive/ka-agent 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/ka-agent/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive/ka-agent 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-graph/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-graph 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-graph/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-graph 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/main.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/perf_core.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/perf_core.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/em_default_v4.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/em_default_v6.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/em_route_parse.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_acl.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_acl.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_acl_scalar.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_altivec.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em_hlm.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em_hlm_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em_hlm_sse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em_sequential.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_event.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_event.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_event_generic.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_event_internal_port.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_fib.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm_altivec.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm_sse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:58.851 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_route.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_sse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/lpm_default_v4.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/lpm_default_v6.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/lpm_route_parse.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/link_status_interrupt/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/link_status_interrupt 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/link_status_interrupt/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/link_status_interrupt 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_client/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_client 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_client/client.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_client 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/args.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/args.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/init.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/init.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/shared/common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/shared 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/hotplug_mp/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/hotplug_mp 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/hotplug_mp/commands.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/hotplug_mp 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/hotplug_mp/commands.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/hotplug_mp 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/hotplug_mp/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/hotplug_mp 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/simple_mp/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/simple_mp 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/simple_mp/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/simple_mp 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/simple_mp/mp_commands.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/simple_mp 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/simple_mp/mp_commands.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/simple_mp 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/symmetric_mp/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/symmetric_mp 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/symmetric_mp/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/symmetric_mp 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/ntb/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ntb 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/ntb/ntb_fwd.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ntb 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/packet_ordering/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/packet_ordering 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/packet_ordering/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/packet_ordering 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/cli.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/cli.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/conn.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/conn.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/obj.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/obj.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/thread.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/thread.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/ethdev.io to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib_nexthop_group_table.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib_nexthop_table.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib_routing_table.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/hash_func.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/hash_func.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd_macswp.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd_macswp.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd_macswp_pcap.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd_pcap.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/learner.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/learner.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/meter.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/meter.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/mirroring.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/mirroring.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/packet.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/pcap.io to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/recirculation.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/recirculation.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/registers.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/registers.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/selector.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/selector.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/selector.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/varbit.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/varbit.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan_pcap.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan_table.py to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan_table.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/ptpclient/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ptpclient 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/ptpclient/ptpclient.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ptpclient 00:03:58.852 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/main.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/rte_policer.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/rte_policer.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/app_thread.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/args.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/cfg_file.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/cfg_file.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/cmdline.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/init.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/main.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/profile.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/profile_ov.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/profile_pie.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/profile_red.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/stats.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/rxtx_callbacks/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/rxtx_callbacks 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/rxtx_callbacks/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/rxtx_callbacks 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/node/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/node 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/node/node.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/node 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/server/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/server 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/server/args.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/server 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/server/args.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/server 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/server/init.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/server 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/server/init.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/server 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/server/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/server 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/shared/common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/shared 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/service_cores/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/service_cores 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/service_cores/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/service_cores 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/skeleton/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/skeleton 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/skeleton/basicfwd.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/skeleton 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/timer/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/timer 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/timer/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/timer 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/vdpa/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vdpa 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/vdpa/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vdpa 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/vdpa/vdpa_blk_compact.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vdpa 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost/main.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost/virtio_net.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/blk.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/blk_spec.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/vhost_blk.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/vhost_blk.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/vhost_blk_compat.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_crypto/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_crypto 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_crypto/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_crypto 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/channel_manager.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/channel_manager.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/channel_monitor.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/channel_monitor.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/oob_monitor.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/oob_monitor_nop.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/oob_monitor_x86.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/parse.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/parse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/power_manager.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/power_manager.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/vm_power_cli.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/vm_power_cli.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/parse.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/parse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/vm_power_cli_guest.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/vm_power_cli_guest.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/vmdq/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vmdq 00:03:58.853 Installing /home/vagrant/spdk_repo/dpdk/examples/vmdq/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vmdq 00:03:58.854 Installing /home/vagrant/spdk_repo/dpdk/examples/vmdq_dcb/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vmdq_dcb 00:03:58.854 Installing /home/vagrant/spdk_repo/dpdk/examples/vmdq_dcb/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vmdq_dcb 00:03:58.854 Installing lib/librte_kvargs.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:58.854 Installing lib/librte_kvargs.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:58.854 Installing lib/librte_telemetry.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:58.854 Installing lib/librte_telemetry.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:58.854 Installing lib/librte_eal.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:58.854 Installing lib/librte_eal.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:58.854 Installing lib/librte_ring.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:58.854 Installing lib/librte_ring.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:58.854 Installing lib/librte_rcu.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:58.854 Installing lib/librte_rcu.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:58.854 Installing lib/librte_mempool.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:58.854 Installing lib/librte_mempool.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:58.854 Installing lib/librte_mbuf.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:58.854 Installing lib/librte_mbuf.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:58.854 Installing lib/librte_net.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_net.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_meter.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_meter.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_ethdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_ethdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_pci.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_pci.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_cmdline.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_cmdline.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_metrics.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_metrics.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_hash.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_hash.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_timer.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_timer.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_acl.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_acl.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_bbdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_bbdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_bitratestats.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_bitratestats.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_bpf.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_bpf.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_cfgfile.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_cfgfile.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_compressdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_compressdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_cryptodev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_cryptodev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_distributor.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_distributor.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_efd.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_efd.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_eventdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_eventdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_gpudev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_gpudev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_gro.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_gro.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_gso.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_gso.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_ip_frag.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_ip_frag.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_jobstats.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_jobstats.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_latencystats.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_latencystats.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_lpm.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_lpm.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_member.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_member.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_pcapng.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_pcapng.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_power.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_power.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_rawdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_rawdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_regexdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_regexdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_dmadev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_dmadev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_rib.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_rib.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_reorder.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_reorder.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_sched.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_sched.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_security.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_security.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_stack.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_stack.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_vhost.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_vhost.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_ipsec.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_ipsec.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_fib.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_fib.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_port.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_port.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_pdump.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_pdump.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_table.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_table.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_pipeline.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_pipeline.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_graph.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_graph.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_node.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing lib/librte_node.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing drivers/librte_bus_pci.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing drivers/librte_bus_pci.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0 00:03:59.115 Installing drivers/librte_bus_vdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing drivers/librte_bus_vdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0 00:03:59.115 Installing drivers/librte_mempool_ring.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing drivers/librte_mempool_ring.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0 00:03:59.115 Installing drivers/librte_net_i40e.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.115 Installing drivers/librte_net_i40e.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0 00:03:59.115 Installing app/dpdk-dumpcap to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:59.115 Installing app/dpdk-pdump to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:59.115 Installing app/dpdk-proc-info to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:59.116 Installing app/dpdk-test-acl to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:59.116 Installing app/dpdk-test-bbdev to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:59.116 Installing app/dpdk-test-cmdline to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:59.116 Installing app/dpdk-test-compress-perf to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:59.116 Installing app/dpdk-test-crypto-perf to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:59.116 Installing app/dpdk-test-eventdev to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:59.116 Installing app/dpdk-test-fib to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:59.116 Installing app/dpdk-test-flow-perf to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:59.116 Installing app/dpdk-test-gpudev to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:59.116 Installing app/dpdk-test-pipeline to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:59.116 Installing app/dpdk-testpmd to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:59.116 Installing app/dpdk-test-regex to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:59.116 Installing app/dpdk-test-sad to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:59.116 Installing app/dpdk-test-security-perf to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/config/rte_config.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/kvargs/rte_kvargs.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/telemetry/rte_telemetry.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_atomic.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_byteorder.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_cpuflags.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_cycles.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_io.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_memcpy.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_pause.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_power_intrinsics.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_prefetch.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_rwlock.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_spinlock.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_vect.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_atomic.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_byteorder.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_cpuflags.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_cycles.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_io.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_memcpy.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_pause.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_power_intrinsics.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_prefetch.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_rtm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_rwlock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_spinlock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_vect.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_atomic_32.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_atomic_64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_byteorder_32.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_byteorder_64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_alarm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_bitmap.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_bitops.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_branch_prediction.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_bus.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_class.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_common.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_compat.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_debug.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_dev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_devargs.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_eal.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_eal_memconfig.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_eal_trace.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_errno.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_epoll.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_fbarray.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_hexdump.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_hypervisor.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_interrupts.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_keepalive.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_launch.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_lcore.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_log.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_malloc.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_mcslock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_memory.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_memzone.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_pci_dev_feature_defs.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_pci_dev_features.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_per_lcore.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_pflock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_random.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_reciprocal.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_seqcount.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_seqlock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_service.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_service_component.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_string_fns.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_tailq.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_thread.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_ticketlock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_time.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_trace.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_trace_point.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_trace_point_register.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_uuid.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_version.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_vfio.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/linux/include/rte_os.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_elem.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_elem_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_c11_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_generic_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_hts.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_hts_elem_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.116 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_peek.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_peek_elem_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_peek_zc.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_rts.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_rts_elem_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/rcu/rte_rcu_qsbr.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/mempool/rte_mempool.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/mempool/rte_mempool_trace.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/mempool/rte_mempool_trace_fp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf_ptype.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf_pool_ops.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf_dyn.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_ip.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_tcp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_udp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_esp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_sctp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_icmp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_arp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_ether.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_macsec.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_vxlan.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_gre.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_gtp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_net.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_net_crc.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_mpls.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_higig.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_ecpri.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_geneve.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_l2tpv2.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_ppp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/meter/rte_meter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_cman.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_ethdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_ethdev_trace.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_ethdev_trace_fp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_dev_info.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_flow.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_flow_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_mtr.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_mtr_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_tm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_tm_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_ethdev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_eth_ctrl.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/pci/rte_pci.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_num.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_ipaddr.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_etheraddr.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_string.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_rdline.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_vt100.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_socket.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_cirbuf.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_portlist.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/metrics/rte_metrics.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/metrics/rte_metrics_telemetry.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_fbk_hash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_hash_crc.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_hash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_jhash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_thash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_thash_gfni.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_crc_arm64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_crc_generic.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_crc_sw.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_crc_x86.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_thash_x86_gfni.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/timer/rte_timer.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/acl/rte_acl.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/acl/rte_acl_osdep.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/bbdev/rte_bbdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/bbdev/rte_bbdev_pmd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/bbdev/rte_bbdev_op.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/bitratestats/rte_bitrate.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/bpf/bpf_def.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/bpf/rte_bpf.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/bpf/rte_bpf_ethdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/cfgfile/rte_cfgfile.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/compressdev/rte_compressdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/compressdev/rte_comp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_cryptodev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_cryptodev_trace.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_cryptodev_trace_fp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_crypto.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_crypto_sym.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_crypto_asym.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_cryptodev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/distributor/rte_distributor.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/efd/rte_efd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_crypto_adapter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_eth_rx_adapter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_eth_tx_adapter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_ring.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_timer_adapter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_eventdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_eventdev_trace_fp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_eventdev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/gpudev/rte_gpudev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/gro/rte_gro.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/gso/rte_gso.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.117 Installing /home/vagrant/spdk_repo/dpdk/lib/ip_frag/rte_ip_frag.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/jobstats/rte_jobstats.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/latencystats/rte_latencystats.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm6.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_altivec.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_neon.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_scalar.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_sse.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_sve.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/member/rte_member.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/pcapng/rte_pcapng.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/power/rte_power.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/power/rte_power_empty_poll.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/power/rte_power_intel_uncore.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/power/rte_power_pmd_mgmt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/power/rte_power_guest_channel.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/rawdev/rte_rawdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/rawdev/rte_rawdev_pmd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/regexdev/rte_regexdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/regexdev/rte_regexdev_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/regexdev/rte_regexdev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/dmadev/rte_dmadev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/dmadev/rte_dmadev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/rib/rte_rib.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/rib/rte_rib6.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/reorder/rte_reorder.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_approx.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_red.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_sched.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_sched_common.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_pie.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/security/rte_security.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/security/rte_security_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_std.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_lf.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_lf_generic.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_lf_c11.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_lf_stubs.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/vhost/rte_vdpa.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/vhost/rte_vhost.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/vhost/rte_vhost_async.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/vhost/rte_vhost_crypto.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/ipsec/rte_ipsec.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/ipsec/rte_ipsec_sa.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/ipsec/rte_ipsec_sad.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/ipsec/rte_ipsec_group.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/fib/rte_fib.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/fib/rte_fib6.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_ethdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_fd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_frag.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_ras.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_ring.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_sched.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_source_sink.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_sym_crypto.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_eventdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port_ethdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port_fd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port_ring.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port_source_sink.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/pdump/rte_pdump.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_lru.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_hash_func.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table_em.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table_learner.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table_selector.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table_wm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_acl.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_array.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_hash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_hash_cuckoo.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_hash_func.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_lpm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_lpm_ipv6.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_stub.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_lru_arm64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_lru_x86.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_hash_func_arm64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_pipeline.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_port_in_action.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_table_action.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_swx_pipeline.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_swx_extern.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_swx_ctl.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/graph/rte_graph.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/graph/rte_graph_worker.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/node/rte_node_ip4_api.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/lib/node/rte_node_eth_api.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/drivers/bus/pci/rte_bus_pci.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/drivers/bus/vdev/rte_bus_vdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/drivers/net/i40e/rte_pmd_i40e.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/usertools/dpdk-devbind.py to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/usertools/dpdk-pmdinfo.py to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/usertools/dpdk-telemetry.py to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/usertools/dpdk-hugepages.py to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/build-tmp/rte_build_config.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/build-tmp/meson-private/libdpdk-libs.pc to /home/vagrant/spdk_repo/dpdk/build/lib/pkgconfig 00:03:59.118 Installing /home/vagrant/spdk_repo/dpdk/build-tmp/meson-private/libdpdk.pc to /home/vagrant/spdk_repo/dpdk/build/lib/pkgconfig 00:03:59.119 Installing symlink pointing to librte_kvargs.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_kvargs.so.23 00:03:59.119 Installing symlink pointing to librte_kvargs.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_kvargs.so 00:03:59.119 Installing symlink pointing to librte_telemetry.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_telemetry.so.23 00:03:59.119 Installing symlink pointing to librte_telemetry.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_telemetry.so 00:03:59.119 Installing symlink pointing to librte_eal.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_eal.so.23 00:03:59.119 Installing symlink pointing to librte_eal.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_eal.so 00:03:59.119 Installing symlink pointing to librte_ring.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ring.so.23 00:03:59.119 Installing symlink pointing to librte_ring.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ring.so 00:03:59.119 Installing symlink pointing to librte_rcu.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rcu.so.23 00:03:59.119 Installing symlink pointing to librte_rcu.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rcu.so 00:03:59.119 Installing symlink pointing to librte_mempool.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_mempool.so.23 00:03:59.119 Installing symlink pointing to librte_mempool.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_mempool.so 00:03:59.119 Installing symlink pointing to librte_mbuf.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_mbuf.so.23 00:03:59.119 Installing symlink pointing to librte_mbuf.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_mbuf.so 00:03:59.119 Installing symlink pointing to librte_net.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_net.so.23 00:03:59.119 Installing symlink pointing to librte_net.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_net.so 00:03:59.119 Installing symlink pointing to librte_meter.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_meter.so.23 00:03:59.119 Installing symlink pointing to librte_meter.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_meter.so 00:03:59.119 Installing symlink pointing to librte_ethdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ethdev.so.23 00:03:59.119 Installing symlink pointing to librte_ethdev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ethdev.so 00:03:59.119 Installing symlink pointing to librte_pci.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pci.so.23 00:03:59.119 Installing symlink pointing to librte_pci.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pci.so 00:03:59.119 Installing symlink pointing to librte_cmdline.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cmdline.so.23 00:03:59.119 Installing symlink pointing to librte_cmdline.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cmdline.so 00:03:59.119 Installing symlink pointing to librte_metrics.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_metrics.so.23 00:03:59.119 Installing symlink pointing to librte_metrics.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_metrics.so 00:03:59.119 Installing symlink pointing to librte_hash.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_hash.so.23 00:03:59.119 Installing symlink pointing to librte_hash.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_hash.so 00:03:59.119 Installing symlink pointing to librte_timer.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_timer.so.23 00:03:59.119 Installing symlink pointing to librte_timer.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_timer.so 00:03:59.119 Installing symlink pointing to librte_acl.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_acl.so.23 00:03:59.119 Installing symlink pointing to librte_acl.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_acl.so 00:03:59.119 Installing symlink pointing to librte_bbdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bbdev.so.23 00:03:59.119 Installing symlink pointing to librte_bbdev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bbdev.so 00:03:59.119 Installing symlink pointing to librte_bitratestats.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bitratestats.so.23 00:03:59.119 Installing symlink pointing to librte_bitratestats.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bitratestats.so 00:03:59.119 Installing symlink pointing to librte_bpf.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bpf.so.23 00:03:59.119 Installing symlink pointing to librte_bpf.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bpf.so 00:03:59.119 Installing symlink pointing to librte_cfgfile.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cfgfile.so.23 00:03:59.119 Installing symlink pointing to librte_cfgfile.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cfgfile.so 00:03:59.119 Installing symlink pointing to librte_compressdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_compressdev.so.23 00:03:59.119 Installing symlink pointing to librte_compressdev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_compressdev.so 00:03:59.119 Installing symlink pointing to librte_cryptodev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cryptodev.so.23 00:03:59.119 Installing symlink pointing to librte_cryptodev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cryptodev.so 00:03:59.119 Installing symlink pointing to librte_distributor.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_distributor.so.23 00:03:59.119 Installing symlink pointing to librte_distributor.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_distributor.so 00:03:59.119 Installing symlink pointing to librte_efd.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_efd.so.23 00:03:59.119 Installing symlink pointing to librte_efd.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_efd.so 00:03:59.119 Installing symlink pointing to librte_eventdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_eventdev.so.23 00:03:59.119 Installing symlink pointing to librte_eventdev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_eventdev.so 00:03:59.119 Installing symlink pointing to librte_gpudev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gpudev.so.23 00:03:59.379 './librte_bus_pci.so' -> 'dpdk/pmds-23.0/librte_bus_pci.so' 00:03:59.379 './librte_bus_pci.so.23' -> 'dpdk/pmds-23.0/librte_bus_pci.so.23' 00:03:59.379 './librte_bus_pci.so.23.0' -> 'dpdk/pmds-23.0/librte_bus_pci.so.23.0' 00:03:59.379 './librte_bus_vdev.so' -> 'dpdk/pmds-23.0/librte_bus_vdev.so' 00:03:59.379 './librte_bus_vdev.so.23' -> 'dpdk/pmds-23.0/librte_bus_vdev.so.23' 00:03:59.379 './librte_bus_vdev.so.23.0' -> 'dpdk/pmds-23.0/librte_bus_vdev.so.23.0' 00:03:59.379 './librte_mempool_ring.so' -> 'dpdk/pmds-23.0/librte_mempool_ring.so' 00:03:59.379 './librte_mempool_ring.so.23' -> 'dpdk/pmds-23.0/librte_mempool_ring.so.23' 00:03:59.379 './librte_mempool_ring.so.23.0' -> 'dpdk/pmds-23.0/librte_mempool_ring.so.23.0' 00:03:59.379 './librte_net_i40e.so' -> 'dpdk/pmds-23.0/librte_net_i40e.so' 00:03:59.379 './librte_net_i40e.so.23' -> 'dpdk/pmds-23.0/librte_net_i40e.so.23' 00:03:59.379 './librte_net_i40e.so.23.0' -> 'dpdk/pmds-23.0/librte_net_i40e.so.23.0' 00:03:59.379 Installing symlink pointing to librte_gpudev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gpudev.so 00:03:59.379 Installing symlink pointing to librte_gro.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gro.so.23 00:03:59.379 Installing symlink pointing to librte_gro.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gro.so 00:03:59.379 Installing symlink pointing to librte_gso.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gso.so.23 00:03:59.379 Installing symlink pointing to librte_gso.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gso.so 00:03:59.379 Installing symlink pointing to librte_ip_frag.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ip_frag.so.23 00:03:59.379 Installing symlink pointing to librte_ip_frag.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ip_frag.so 00:03:59.379 Installing symlink pointing to librte_jobstats.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_jobstats.so.23 00:03:59.379 Installing symlink pointing to librte_jobstats.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_jobstats.so 00:03:59.379 Installing symlink pointing to librte_latencystats.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_latencystats.so.23 00:03:59.379 Installing symlink pointing to librte_latencystats.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_latencystats.so 00:03:59.379 Installing symlink pointing to librte_lpm.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_lpm.so.23 00:03:59.379 Installing symlink pointing to librte_lpm.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_lpm.so 00:03:59.379 Installing symlink pointing to librte_member.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_member.so.23 00:03:59.379 Installing symlink pointing to librte_member.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_member.so 00:03:59.379 Installing symlink pointing to librte_pcapng.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pcapng.so.23 00:03:59.379 Installing symlink pointing to librte_pcapng.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pcapng.so 00:03:59.379 Installing symlink pointing to librte_power.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_power.so.23 00:03:59.379 Installing symlink pointing to librte_power.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_power.so 00:03:59.379 Installing symlink pointing to librte_rawdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rawdev.so.23 00:03:59.379 Installing symlink pointing to librte_rawdev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rawdev.so 00:03:59.379 Installing symlink pointing to librte_regexdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_regexdev.so.23 00:03:59.379 Installing symlink pointing to librte_regexdev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_regexdev.so 00:03:59.379 Installing symlink pointing to librte_dmadev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_dmadev.so.23 00:03:59.379 Installing symlink pointing to librte_dmadev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_dmadev.so 00:03:59.379 Installing symlink pointing to librte_rib.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rib.so.23 00:03:59.379 Installing symlink pointing to librte_rib.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rib.so 00:03:59.379 Installing symlink pointing to librte_reorder.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_reorder.so.23 00:03:59.379 Installing symlink pointing to librte_reorder.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_reorder.so 00:03:59.379 Installing symlink pointing to librte_sched.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_sched.so.23 00:03:59.379 Installing symlink pointing to librte_sched.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_sched.so 00:03:59.379 Installing symlink pointing to librte_security.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_security.so.23 00:03:59.379 Installing symlink pointing to librte_security.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_security.so 00:03:59.379 Installing symlink pointing to librte_stack.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_stack.so.23 00:03:59.379 Installing symlink pointing to librte_stack.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_stack.so 00:03:59.379 Installing symlink pointing to librte_vhost.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_vhost.so.23 00:03:59.379 Installing symlink pointing to librte_vhost.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_vhost.so 00:03:59.379 Installing symlink pointing to librte_ipsec.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ipsec.so.23 00:03:59.379 Installing symlink pointing to librte_ipsec.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ipsec.so 00:03:59.379 Installing symlink pointing to librte_fib.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_fib.so.23 00:03:59.379 Installing symlink pointing to librte_fib.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_fib.so 00:03:59.379 Installing symlink pointing to librte_port.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_port.so.23 00:03:59.379 Installing symlink pointing to librte_port.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_port.so 00:03:59.379 Installing symlink pointing to librte_pdump.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pdump.so.23 00:03:59.379 Installing symlink pointing to librte_pdump.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pdump.so 00:03:59.379 Installing symlink pointing to librte_table.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_table.so.23 00:03:59.379 Installing symlink pointing to librte_table.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_table.so 00:03:59.379 Installing symlink pointing to librte_pipeline.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pipeline.so.23 00:03:59.379 Installing symlink pointing to librte_pipeline.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pipeline.so 00:03:59.379 Installing symlink pointing to librte_graph.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_graph.so.23 00:03:59.379 Installing symlink pointing to librte_graph.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_graph.so 00:03:59.379 Installing symlink pointing to librte_node.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_node.so.23 00:03:59.379 Installing symlink pointing to librte_node.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_node.so 00:03:59.379 Installing symlink pointing to librte_bus_pci.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_pci.so.23 00:03:59.379 Installing symlink pointing to librte_bus_pci.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_pci.so 00:03:59.379 Installing symlink pointing to librte_bus_vdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_vdev.so.23 00:03:59.379 Installing symlink pointing to librte_bus_vdev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_vdev.so 00:03:59.379 Installing symlink pointing to librte_mempool_ring.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_mempool_ring.so.23 00:03:59.379 Installing symlink pointing to librte_mempool_ring.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_mempool_ring.so 00:03:59.379 Installing symlink pointing to librte_net_i40e.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_net_i40e.so.23 00:03:59.379 Installing symlink pointing to librte_net_i40e.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_net_i40e.so 00:03:59.379 Running custom install script '/bin/sh /home/vagrant/spdk_repo/dpdk/config/../buildtools/symlink-drivers-solibs.sh lib dpdk/pmds-23.0' 00:03:59.379 15:09:49 build_native_dpdk -- common/autobuild_common.sh@213 -- $ cat 00:03:59.379 15:09:49 build_native_dpdk -- common/autobuild_common.sh@218 -- $ cd /home/vagrant/spdk_repo/spdk 00:03:59.379 00:03:59.379 real 0m47.921s 00:03:59.379 user 4m33.516s 00:03:59.379 sys 0m56.405s 00:03:59.379 15:09:49 build_native_dpdk -- common/autotest_common.sh@1130 -- $ xtrace_disable 00:03:59.379 15:09:49 build_native_dpdk -- common/autotest_common.sh@10 -- $ set +x 00:03:59.379 ************************************ 00:03:59.379 END TEST build_native_dpdk 00:03:59.379 ************************************ 00:03:59.379 15:09:49 -- spdk/autobuild.sh@31 -- $ case "$SPDK_TEST_AUTOBUILD" in 00:03:59.379 15:09:49 -- spdk/autobuild.sh@47 -- $ [[ 0 -eq 1 ]] 00:03:59.379 15:09:49 -- spdk/autobuild.sh@51 -- $ [[ 0 -eq 1 ]] 00:03:59.379 15:09:49 -- spdk/autobuild.sh@55 -- $ [[ -n '' ]] 00:03:59.379 15:09:49 -- spdk/autobuild.sh@57 -- $ [[ 0 -eq 1 ]] 00:03:59.379 15:09:49 -- spdk/autobuild.sh@59 -- $ [[ 0 -eq 1 ]] 00:03:59.379 15:09:49 -- spdk/autobuild.sh@62 -- $ [[ 0 -eq 1 ]] 00:03:59.379 15:09:49 -- spdk/autobuild.sh@67 -- $ /home/vagrant/spdk_repo/spdk/configure --enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --enable-ubsan --enable-asan --enable-coverage --with-ublk --with-raid5f --with-dpdk=/home/vagrant/spdk_repo/dpdk/build --with-shared 00:03:59.639 Using /home/vagrant/spdk_repo/dpdk/build/lib/pkgconfig for additional libs... 00:03:59.639 DPDK libraries: /home/vagrant/spdk_repo/dpdk/build/lib 00:03:59.639 DPDK includes: //home/vagrant/spdk_repo/dpdk/build/include 00:03:59.639 Using default SPDK env in /home/vagrant/spdk_repo/spdk/lib/env_dpdk 00:03:59.897 Using 'verbs' RDMA provider 00:04:16.207 Configuring ISA-L (logfile: /home/vagrant/spdk_repo/spdk/.spdk-isal.log)...done. 00:04:34.309 Configuring ISA-L-crypto (logfile: /home/vagrant/spdk_repo/spdk/.spdk-isal-crypto.log)...done. 00:04:34.309 Creating mk/config.mk...done. 00:04:34.309 Creating mk/cc.flags.mk...done. 00:04:34.309 Type 'make' to build. 00:04:34.309 15:10:23 -- spdk/autobuild.sh@70 -- $ run_test make make -j10 00:04:34.309 15:10:23 -- common/autotest_common.sh@1105 -- $ '[' 3 -le 1 ']' 00:04:34.309 15:10:23 -- common/autotest_common.sh@1111 -- $ xtrace_disable 00:04:34.309 15:10:23 -- common/autotest_common.sh@10 -- $ set +x 00:04:34.309 ************************************ 00:04:34.309 START TEST make 00:04:34.309 ************************************ 00:04:34.309 15:10:23 make -- common/autotest_common.sh@1129 -- $ make -j10 00:04:34.309 make[1]: Nothing to be done for 'all'. 00:05:21.009 CC lib/log/log.o 00:05:21.009 CC lib/log/log_flags.o 00:05:21.009 CC lib/ut/ut.o 00:05:21.009 CC lib/log/log_deprecated.o 00:05:21.009 CC lib/ut_mock/mock.o 00:05:21.009 LIB libspdk_log.a 00:05:21.009 LIB libspdk_ut.a 00:05:21.009 LIB libspdk_ut_mock.a 00:05:21.009 SO libspdk_log.so.7.1 00:05:21.009 SO libspdk_ut.so.2.0 00:05:21.009 SO libspdk_ut_mock.so.6.0 00:05:21.009 SYMLINK libspdk_ut.so 00:05:21.009 SYMLINK libspdk_log.so 00:05:21.009 SYMLINK libspdk_ut_mock.so 00:05:21.009 CC lib/dma/dma.o 00:05:21.009 CC lib/ioat/ioat.o 00:05:21.009 CC lib/util/base64.o 00:05:21.009 CC lib/util/bit_array.o 00:05:21.009 CC lib/util/crc32c.o 00:05:21.009 CC lib/util/cpuset.o 00:05:21.009 CC lib/util/crc16.o 00:05:21.009 CC lib/util/crc32.o 00:05:21.009 CXX lib/trace_parser/trace.o 00:05:21.009 CC lib/vfio_user/host/vfio_user_pci.o 00:05:21.009 CC lib/util/crc32_ieee.o 00:05:21.009 CC lib/util/crc64.o 00:05:21.009 CC lib/vfio_user/host/vfio_user.o 00:05:21.009 LIB libspdk_dma.a 00:05:21.009 CC lib/util/dif.o 00:05:21.009 SO libspdk_dma.so.5.0 00:05:21.009 CC lib/util/fd.o 00:05:21.009 CC lib/util/fd_group.o 00:05:21.009 SYMLINK libspdk_dma.so 00:05:21.009 CC lib/util/file.o 00:05:21.009 CC lib/util/hexlify.o 00:05:21.009 CC lib/util/iov.o 00:05:21.009 LIB libspdk_ioat.a 00:05:21.009 SO libspdk_ioat.so.7.0 00:05:21.009 CC lib/util/math.o 00:05:21.009 CC lib/util/net.o 00:05:21.009 LIB libspdk_vfio_user.a 00:05:21.009 SYMLINK libspdk_ioat.so 00:05:21.268 CC lib/util/pipe.o 00:05:21.268 CC lib/util/strerror_tls.o 00:05:21.268 SO libspdk_vfio_user.so.5.0 00:05:21.268 CC lib/util/string.o 00:05:21.268 SYMLINK libspdk_vfio_user.so 00:05:21.268 CC lib/util/uuid.o 00:05:21.268 CC lib/util/xor.o 00:05:21.268 CC lib/util/zipf.o 00:05:21.268 CC lib/util/md5.o 00:05:21.528 LIB libspdk_util.a 00:05:21.788 SO libspdk_util.so.10.1 00:05:21.788 LIB libspdk_trace_parser.a 00:05:21.788 SYMLINK libspdk_util.so 00:05:21.788 SO libspdk_trace_parser.so.6.0 00:05:22.048 SYMLINK libspdk_trace_parser.so 00:05:22.048 CC lib/env_dpdk/memory.o 00:05:22.048 CC lib/env_dpdk/env.o 00:05:22.048 CC lib/env_dpdk/pci.o 00:05:22.048 CC lib/env_dpdk/threads.o 00:05:22.048 CC lib/env_dpdk/init.o 00:05:22.048 CC lib/vmd/vmd.o 00:05:22.048 CC lib/idxd/idxd.o 00:05:22.048 CC lib/json/json_parse.o 00:05:22.048 CC lib/rdma_utils/rdma_utils.o 00:05:22.048 CC lib/conf/conf.o 00:05:22.048 CC lib/env_dpdk/pci_ioat.o 00:05:22.308 LIB libspdk_conf.a 00:05:22.308 CC lib/json/json_util.o 00:05:22.308 CC lib/json/json_write.o 00:05:22.308 SO libspdk_conf.so.6.0 00:05:22.308 LIB libspdk_rdma_utils.a 00:05:22.308 SO libspdk_rdma_utils.so.1.0 00:05:22.308 SYMLINK libspdk_conf.so 00:05:22.308 CC lib/env_dpdk/pci_virtio.o 00:05:22.308 CC lib/idxd/idxd_user.o 00:05:22.308 CC lib/env_dpdk/pci_vmd.o 00:05:22.308 SYMLINK libspdk_rdma_utils.so 00:05:22.308 CC lib/env_dpdk/pci_idxd.o 00:05:22.569 CC lib/vmd/led.o 00:05:22.569 CC lib/idxd/idxd_kernel.o 00:05:22.569 CC lib/env_dpdk/pci_event.o 00:05:22.569 CC lib/env_dpdk/sigbus_handler.o 00:05:22.569 CC lib/rdma_provider/common.o 00:05:22.569 LIB libspdk_json.a 00:05:22.569 SO libspdk_json.so.6.0 00:05:22.569 CC lib/env_dpdk/pci_dpdk.o 00:05:22.569 CC lib/rdma_provider/rdma_provider_verbs.o 00:05:22.569 SYMLINK libspdk_json.so 00:05:22.569 CC lib/env_dpdk/pci_dpdk_2207.o 00:05:22.569 CC lib/env_dpdk/pci_dpdk_2211.o 00:05:22.829 LIB libspdk_idxd.a 00:05:22.829 LIB libspdk_vmd.a 00:05:22.829 SO libspdk_idxd.so.12.1 00:05:22.829 SO libspdk_vmd.so.6.0 00:05:22.829 SYMLINK libspdk_idxd.so 00:05:22.829 SYMLINK libspdk_vmd.so 00:05:22.829 CC lib/jsonrpc/jsonrpc_server.o 00:05:22.829 CC lib/jsonrpc/jsonrpc_client_tcp.o 00:05:22.829 CC lib/jsonrpc/jsonrpc_client.o 00:05:22.829 CC lib/jsonrpc/jsonrpc_server_tcp.o 00:05:22.829 LIB libspdk_rdma_provider.a 00:05:22.829 SO libspdk_rdma_provider.so.7.0 00:05:22.829 SYMLINK libspdk_rdma_provider.so 00:05:23.088 LIB libspdk_jsonrpc.a 00:05:23.088 SO libspdk_jsonrpc.so.6.0 00:05:23.347 SYMLINK libspdk_jsonrpc.so 00:05:23.606 LIB libspdk_env_dpdk.a 00:05:23.606 CC lib/rpc/rpc.o 00:05:23.606 SO libspdk_env_dpdk.so.15.1 00:05:23.865 SYMLINK libspdk_env_dpdk.so 00:05:23.865 LIB libspdk_rpc.a 00:05:23.865 SO libspdk_rpc.so.6.0 00:05:23.865 SYMLINK libspdk_rpc.so 00:05:24.432 CC lib/trace/trace.o 00:05:24.432 CC lib/trace/trace_rpc.o 00:05:24.432 CC lib/trace/trace_flags.o 00:05:24.432 CC lib/keyring/keyring.o 00:05:24.432 CC lib/keyring/keyring_rpc.o 00:05:24.432 CC lib/notify/notify.o 00:05:24.432 CC lib/notify/notify_rpc.o 00:05:24.432 LIB libspdk_notify.a 00:05:24.691 SO libspdk_notify.so.6.0 00:05:24.691 LIB libspdk_trace.a 00:05:24.691 LIB libspdk_keyring.a 00:05:24.691 SYMLINK libspdk_notify.so 00:05:24.691 SO libspdk_keyring.so.2.0 00:05:24.691 SO libspdk_trace.so.11.0 00:05:24.691 SYMLINK libspdk_keyring.so 00:05:24.691 SYMLINK libspdk_trace.so 00:05:25.258 CC lib/sock/sock.o 00:05:25.258 CC lib/sock/sock_rpc.o 00:05:25.258 CC lib/thread/thread.o 00:05:25.258 CC lib/thread/iobuf.o 00:05:25.518 LIB libspdk_sock.a 00:05:25.777 SO libspdk_sock.so.10.0 00:05:25.777 SYMLINK libspdk_sock.so 00:05:26.035 CC lib/nvme/nvme_ctrlr.o 00:05:26.035 CC lib/nvme/nvme_ctrlr_cmd.o 00:05:26.035 CC lib/nvme/nvme_fabric.o 00:05:26.035 CC lib/nvme/nvme_ns_cmd.o 00:05:26.035 CC lib/nvme/nvme_ns.o 00:05:26.035 CC lib/nvme/nvme_pcie_common.o 00:05:26.035 CC lib/nvme/nvme_pcie.o 00:05:26.035 CC lib/nvme/nvme_qpair.o 00:05:26.035 CC lib/nvme/nvme.o 00:05:27.021 CC lib/nvme/nvme_quirks.o 00:05:27.021 CC lib/nvme/nvme_transport.o 00:05:27.021 LIB libspdk_thread.a 00:05:27.021 CC lib/nvme/nvme_discovery.o 00:05:27.021 SO libspdk_thread.so.11.0 00:05:27.021 CC lib/nvme/nvme_ctrlr_ocssd_cmd.o 00:05:27.021 SYMLINK libspdk_thread.so 00:05:27.021 CC lib/nvme/nvme_ns_ocssd_cmd.o 00:05:27.021 CC lib/nvme/nvme_tcp.o 00:05:27.021 CC lib/nvme/nvme_opal.o 00:05:27.285 CC lib/accel/accel.o 00:05:27.285 CC lib/accel/accel_rpc.o 00:05:27.285 CC lib/accel/accel_sw.o 00:05:27.545 CC lib/nvme/nvme_io_msg.o 00:05:27.804 CC lib/blob/blobstore.o 00:05:27.804 CC lib/init/json_config.o 00:05:27.804 CC lib/blob/request.o 00:05:27.804 CC lib/virtio/virtio.o 00:05:27.804 CC lib/fsdev/fsdev.o 00:05:27.804 CC lib/init/subsystem.o 00:05:27.804 CC lib/init/subsystem_rpc.o 00:05:27.804 CC lib/virtio/virtio_vhost_user.o 00:05:28.064 CC lib/init/rpc.o 00:05:28.064 CC lib/virtio/virtio_vfio_user.o 00:05:28.064 CC lib/virtio/virtio_pci.o 00:05:28.323 LIB libspdk_init.a 00:05:28.323 SO libspdk_init.so.6.0 00:05:28.323 CC lib/fsdev/fsdev_io.o 00:05:28.323 CC lib/fsdev/fsdev_rpc.o 00:05:28.323 SYMLINK libspdk_init.so 00:05:28.323 CC lib/blob/zeroes.o 00:05:28.583 LIB libspdk_virtio.a 00:05:28.583 CC lib/blob/blob_bs_dev.o 00:05:28.583 CC lib/nvme/nvme_poll_group.o 00:05:28.583 LIB libspdk_accel.a 00:05:28.583 SO libspdk_virtio.so.7.0 00:05:28.583 SO libspdk_accel.so.16.0 00:05:28.583 CC lib/nvme/nvme_zns.o 00:05:28.583 CC lib/event/app.o 00:05:28.583 SYMLINK libspdk_virtio.so 00:05:28.583 CC lib/nvme/nvme_stubs.o 00:05:28.583 SYMLINK libspdk_accel.so 00:05:28.583 CC lib/nvme/nvme_auth.o 00:05:28.583 LIB libspdk_fsdev.a 00:05:28.843 SO libspdk_fsdev.so.2.0 00:05:28.843 CC lib/event/reactor.o 00:05:28.843 SYMLINK libspdk_fsdev.so 00:05:28.843 CC lib/nvme/nvme_cuse.o 00:05:28.843 CC lib/event/log_rpc.o 00:05:28.843 CC lib/bdev/bdev.o 00:05:29.102 CC lib/event/app_rpc.o 00:05:29.102 CC lib/event/scheduler_static.o 00:05:29.102 CC lib/bdev/bdev_rpc.o 00:05:29.102 CC lib/bdev/bdev_zone.o 00:05:29.362 CC lib/nvme/nvme_rdma.o 00:05:29.362 CC lib/bdev/part.o 00:05:29.362 CC lib/fuse_dispatcher/fuse_dispatcher.o 00:05:29.362 LIB libspdk_event.a 00:05:29.362 SO libspdk_event.so.14.0 00:05:29.362 CC lib/bdev/scsi_nvme.o 00:05:29.362 SYMLINK libspdk_event.so 00:05:29.931 LIB libspdk_fuse_dispatcher.a 00:05:30.192 SO libspdk_fuse_dispatcher.so.1.0 00:05:30.192 SYMLINK libspdk_fuse_dispatcher.so 00:05:30.761 LIB libspdk_nvme.a 00:05:30.761 SO libspdk_nvme.so.15.0 00:05:31.021 SYMLINK libspdk_nvme.so 00:05:31.591 LIB libspdk_blob.a 00:05:31.591 SO libspdk_blob.so.11.0 00:05:31.591 LIB libspdk_bdev.a 00:05:31.591 SYMLINK libspdk_blob.so 00:05:31.591 SO libspdk_bdev.so.17.0 00:05:31.852 SYMLINK libspdk_bdev.so 00:05:31.852 CC lib/blobfs/blobfs.o 00:05:31.852 CC lib/blobfs/tree.o 00:05:31.852 CC lib/lvol/lvol.o 00:05:31.852 CC lib/ublk/ublk.o 00:05:31.852 CC lib/ublk/ublk_rpc.o 00:05:31.852 CC lib/scsi/dev.o 00:05:31.852 CC lib/scsi/lun.o 00:05:31.852 CC lib/nvmf/ctrlr.o 00:05:31.852 CC lib/nbd/nbd.o 00:05:31.852 CC lib/ftl/ftl_core.o 00:05:32.111 CC lib/ftl/ftl_init.o 00:05:32.111 CC lib/ftl/ftl_layout.o 00:05:32.111 CC lib/ftl/ftl_debug.o 00:05:32.371 CC lib/scsi/port.o 00:05:32.371 CC lib/nvmf/ctrlr_discovery.o 00:05:32.371 CC lib/ftl/ftl_io.o 00:05:32.371 CC lib/ftl/ftl_sb.o 00:05:32.371 CC lib/scsi/scsi.o 00:05:32.371 CC lib/nbd/nbd_rpc.o 00:05:32.371 CC lib/ftl/ftl_l2p.o 00:05:32.630 CC lib/scsi/scsi_bdev.o 00:05:32.630 CC lib/scsi/scsi_pr.o 00:05:32.630 LIB libspdk_nbd.a 00:05:32.630 CC lib/ftl/ftl_l2p_flat.o 00:05:32.630 SO libspdk_nbd.so.7.0 00:05:32.630 LIB libspdk_ublk.a 00:05:32.630 CC lib/scsi/scsi_rpc.o 00:05:32.630 SYMLINK libspdk_nbd.so 00:05:32.630 SO libspdk_ublk.so.3.0 00:05:32.630 CC lib/scsi/task.o 00:05:32.630 SYMLINK libspdk_ublk.so 00:05:32.630 CC lib/nvmf/ctrlr_bdev.o 00:05:32.890 CC lib/nvmf/subsystem.o 00:05:32.890 CC lib/ftl/ftl_nv_cache.o 00:05:32.890 CC lib/nvmf/nvmf.o 00:05:32.890 LIB libspdk_lvol.a 00:05:32.890 LIB libspdk_blobfs.a 00:05:32.890 SO libspdk_lvol.so.10.0 00:05:32.890 SO libspdk_blobfs.so.10.0 00:05:32.890 CC lib/nvmf/nvmf_rpc.o 00:05:32.890 CC lib/nvmf/transport.o 00:05:32.890 SYMLINK libspdk_lvol.so 00:05:32.890 CC lib/ftl/ftl_band.o 00:05:32.890 SYMLINK libspdk_blobfs.so 00:05:32.890 CC lib/ftl/ftl_band_ops.o 00:05:33.150 LIB libspdk_scsi.a 00:05:33.150 SO libspdk_scsi.so.9.0 00:05:33.410 SYMLINK libspdk_scsi.so 00:05:33.410 CC lib/nvmf/tcp.o 00:05:33.410 CC lib/ftl/ftl_writer.o 00:05:33.410 CC lib/ftl/ftl_rq.o 00:05:33.410 CC lib/nvmf/stubs.o 00:05:33.670 CC lib/nvmf/mdns_server.o 00:05:33.670 CC lib/vhost/vhost.o 00:05:33.670 CC lib/iscsi/conn.o 00:05:33.670 CC lib/iscsi/init_grp.o 00:05:33.670 CC lib/nvmf/rdma.o 00:05:33.930 CC lib/nvmf/auth.o 00:05:33.930 CC lib/iscsi/iscsi.o 00:05:33.930 CC lib/ftl/ftl_reloc.o 00:05:33.930 CC lib/iscsi/param.o 00:05:34.190 CC lib/iscsi/portal_grp.o 00:05:34.190 CC lib/iscsi/tgt_node.o 00:05:34.450 CC lib/ftl/ftl_l2p_cache.o 00:05:34.450 CC lib/ftl/ftl_p2l.o 00:05:34.450 CC lib/iscsi/iscsi_subsystem.o 00:05:34.450 CC lib/vhost/vhost_rpc.o 00:05:34.711 CC lib/vhost/vhost_scsi.o 00:05:34.711 CC lib/vhost/vhost_blk.o 00:05:34.711 CC lib/iscsi/iscsi_rpc.o 00:05:34.711 CC lib/ftl/ftl_p2l_log.o 00:05:34.970 CC lib/ftl/mngt/ftl_mngt.o 00:05:34.970 CC lib/vhost/rte_vhost_user.o 00:05:34.970 CC lib/iscsi/task.o 00:05:35.229 CC lib/ftl/mngt/ftl_mngt_bdev.o 00:05:35.229 CC lib/ftl/mngt/ftl_mngt_shutdown.o 00:05:35.229 CC lib/ftl/mngt/ftl_mngt_startup.o 00:05:35.229 CC lib/ftl/mngt/ftl_mngt_md.o 00:05:35.229 CC lib/ftl/mngt/ftl_mngt_misc.o 00:05:35.229 CC lib/ftl/mngt/ftl_mngt_ioch.o 00:05:35.490 CC lib/ftl/mngt/ftl_mngt_l2p.o 00:05:35.490 CC lib/ftl/mngt/ftl_mngt_band.o 00:05:35.490 CC lib/ftl/mngt/ftl_mngt_self_test.o 00:05:35.490 CC lib/ftl/mngt/ftl_mngt_p2l.o 00:05:35.490 LIB libspdk_iscsi.a 00:05:35.490 CC lib/ftl/mngt/ftl_mngt_recovery.o 00:05:35.490 CC lib/ftl/mngt/ftl_mngt_upgrade.o 00:05:35.490 CC lib/ftl/utils/ftl_conf.o 00:05:35.490 CC lib/ftl/utils/ftl_md.o 00:05:35.750 SO libspdk_iscsi.so.8.0 00:05:35.750 CC lib/ftl/utils/ftl_mempool.o 00:05:35.750 CC lib/ftl/utils/ftl_bitmap.o 00:05:35.750 CC lib/ftl/utils/ftl_property.o 00:05:35.750 CC lib/ftl/utils/ftl_layout_tracker_bdev.o 00:05:35.750 CC lib/ftl/upgrade/ftl_layout_upgrade.o 00:05:35.750 SYMLINK libspdk_iscsi.so 00:05:35.750 CC lib/ftl/upgrade/ftl_sb_upgrade.o 00:05:35.750 CC lib/ftl/upgrade/ftl_p2l_upgrade.o 00:05:36.009 LIB libspdk_nvmf.a 00:05:36.009 CC lib/ftl/upgrade/ftl_band_upgrade.o 00:05:36.009 CC lib/ftl/upgrade/ftl_chunk_upgrade.o 00:05:36.009 CC lib/ftl/upgrade/ftl_trim_upgrade.o 00:05:36.009 CC lib/ftl/upgrade/ftl_sb_v3.o 00:05:36.009 SO libspdk_nvmf.so.20.0 00:05:36.009 CC lib/ftl/upgrade/ftl_sb_v5.o 00:05:36.009 CC lib/ftl/nvc/ftl_nvc_dev.o 00:05:36.009 CC lib/ftl/nvc/ftl_nvc_bdev_vss.o 00:05:36.009 LIB libspdk_vhost.a 00:05:36.009 CC lib/ftl/nvc/ftl_nvc_bdev_non_vss.o 00:05:36.010 CC lib/ftl/nvc/ftl_nvc_bdev_common.o 00:05:36.269 SO libspdk_vhost.so.8.0 00:05:36.269 CC lib/ftl/base/ftl_base_dev.o 00:05:36.269 CC lib/ftl/base/ftl_base_bdev.o 00:05:36.269 CC lib/ftl/ftl_trace.o 00:05:36.269 SYMLINK libspdk_vhost.so 00:05:36.269 SYMLINK libspdk_nvmf.so 00:05:36.528 LIB libspdk_ftl.a 00:05:36.787 SO libspdk_ftl.so.9.0 00:05:37.046 SYMLINK libspdk_ftl.so 00:05:37.306 CC module/env_dpdk/env_dpdk_rpc.o 00:05:37.306 CC module/scheduler/dpdk_governor/dpdk_governor.o 00:05:37.306 CC module/accel/error/accel_error.o 00:05:37.306 CC module/keyring/linux/keyring.o 00:05:37.306 CC module/scheduler/gscheduler/gscheduler.o 00:05:37.306 CC module/scheduler/dynamic/scheduler_dynamic.o 00:05:37.306 CC module/sock/posix/posix.o 00:05:37.306 CC module/blob/bdev/blob_bdev.o 00:05:37.566 CC module/fsdev/aio/fsdev_aio.o 00:05:37.566 CC module/keyring/file/keyring.o 00:05:37.566 LIB libspdk_env_dpdk_rpc.a 00:05:37.566 SO libspdk_env_dpdk_rpc.so.6.0 00:05:37.566 LIB libspdk_scheduler_dpdk_governor.a 00:05:37.566 SYMLINK libspdk_env_dpdk_rpc.so 00:05:37.566 CC module/keyring/linux/keyring_rpc.o 00:05:37.566 CC module/keyring/file/keyring_rpc.o 00:05:37.566 CC module/fsdev/aio/fsdev_aio_rpc.o 00:05:37.566 LIB libspdk_scheduler_gscheduler.a 00:05:37.566 SO libspdk_scheduler_dpdk_governor.so.4.0 00:05:37.566 SO libspdk_scheduler_gscheduler.so.4.0 00:05:37.566 LIB libspdk_scheduler_dynamic.a 00:05:37.566 CC module/accel/error/accel_error_rpc.o 00:05:37.566 SYMLINK libspdk_scheduler_dpdk_governor.so 00:05:37.566 SYMLINK libspdk_scheduler_gscheduler.so 00:05:37.566 CC module/fsdev/aio/linux_aio_mgr.o 00:05:37.566 SO libspdk_scheduler_dynamic.so.4.0 00:05:37.566 LIB libspdk_keyring_file.a 00:05:37.826 LIB libspdk_keyring_linux.a 00:05:37.826 SO libspdk_keyring_file.so.2.0 00:05:37.826 LIB libspdk_blob_bdev.a 00:05:37.826 SYMLINK libspdk_scheduler_dynamic.so 00:05:37.826 SO libspdk_keyring_linux.so.1.0 00:05:37.826 SO libspdk_blob_bdev.so.11.0 00:05:37.826 LIB libspdk_accel_error.a 00:05:37.826 SYMLINK libspdk_keyring_file.so 00:05:37.826 CC module/accel/ioat/accel_ioat.o 00:05:37.826 SYMLINK libspdk_keyring_linux.so 00:05:37.826 CC module/accel/ioat/accel_ioat_rpc.o 00:05:37.826 SYMLINK libspdk_blob_bdev.so 00:05:37.826 SO libspdk_accel_error.so.2.0 00:05:37.826 SYMLINK libspdk_accel_error.so 00:05:37.826 CC module/accel/dsa/accel_dsa.o 00:05:37.826 CC module/accel/dsa/accel_dsa_rpc.o 00:05:37.826 CC module/accel/iaa/accel_iaa.o 00:05:37.826 CC module/accel/iaa/accel_iaa_rpc.o 00:05:38.086 LIB libspdk_accel_ioat.a 00:05:38.086 SO libspdk_accel_ioat.so.6.0 00:05:38.086 CC module/bdev/delay/vbdev_delay.o 00:05:38.086 CC module/blobfs/bdev/blobfs_bdev.o 00:05:38.086 LIB libspdk_accel_iaa.a 00:05:38.086 SO libspdk_accel_iaa.so.3.0 00:05:38.086 SYMLINK libspdk_accel_ioat.so 00:05:38.086 CC module/blobfs/bdev/blobfs_bdev_rpc.o 00:05:38.086 CC module/bdev/error/vbdev_error.o 00:05:38.086 SYMLINK libspdk_accel_iaa.so 00:05:38.086 CC module/bdev/error/vbdev_error_rpc.o 00:05:38.086 LIB libspdk_accel_dsa.a 00:05:38.086 CC module/bdev/gpt/gpt.o 00:05:38.086 LIB libspdk_fsdev_aio.a 00:05:38.086 SO libspdk_accel_dsa.so.5.0 00:05:38.086 CC module/bdev/lvol/vbdev_lvol.o 00:05:38.345 CC module/bdev/lvol/vbdev_lvol_rpc.o 00:05:38.346 SO libspdk_fsdev_aio.so.1.0 00:05:38.346 LIB libspdk_sock_posix.a 00:05:38.346 SYMLINK libspdk_accel_dsa.so 00:05:38.346 LIB libspdk_blobfs_bdev.a 00:05:38.346 CC module/bdev/delay/vbdev_delay_rpc.o 00:05:38.346 SYMLINK libspdk_fsdev_aio.so 00:05:38.346 SO libspdk_sock_posix.so.6.0 00:05:38.346 CC module/bdev/gpt/vbdev_gpt.o 00:05:38.346 SO libspdk_blobfs_bdev.so.6.0 00:05:38.346 LIB libspdk_bdev_error.a 00:05:38.346 SYMLINK libspdk_blobfs_bdev.so 00:05:38.346 SYMLINK libspdk_sock_posix.so 00:05:38.346 SO libspdk_bdev_error.so.6.0 00:05:38.346 SYMLINK libspdk_bdev_error.so 00:05:38.346 LIB libspdk_bdev_delay.a 00:05:38.608 SO libspdk_bdev_delay.so.6.0 00:05:38.608 CC module/bdev/malloc/bdev_malloc.o 00:05:38.608 CC module/bdev/null/bdev_null.o 00:05:38.608 CC module/bdev/nvme/bdev_nvme.o 00:05:38.608 CC module/bdev/passthru/vbdev_passthru.o 00:05:38.608 SYMLINK libspdk_bdev_delay.so 00:05:38.608 LIB libspdk_bdev_gpt.a 00:05:38.608 CC module/bdev/raid/bdev_raid.o 00:05:38.608 CC module/bdev/split/vbdev_split.o 00:05:38.608 SO libspdk_bdev_gpt.so.6.0 00:05:38.608 SYMLINK libspdk_bdev_gpt.so 00:05:38.608 CC module/bdev/raid/bdev_raid_rpc.o 00:05:38.608 LIB libspdk_bdev_lvol.a 00:05:38.608 CC module/bdev/zone_block/vbdev_zone_block.o 00:05:38.913 SO libspdk_bdev_lvol.so.6.0 00:05:38.913 CC module/bdev/aio/bdev_aio.o 00:05:38.913 CC module/bdev/null/bdev_null_rpc.o 00:05:38.913 CC module/bdev/passthru/vbdev_passthru_rpc.o 00:05:38.913 SYMLINK libspdk_bdev_lvol.so 00:05:38.913 CC module/bdev/aio/bdev_aio_rpc.o 00:05:38.913 CC module/bdev/split/vbdev_split_rpc.o 00:05:38.913 CC module/bdev/raid/bdev_raid_sb.o 00:05:38.913 CC module/bdev/malloc/bdev_malloc_rpc.o 00:05:38.913 LIB libspdk_bdev_null.a 00:05:38.913 SO libspdk_bdev_null.so.6.0 00:05:38.913 LIB libspdk_bdev_passthru.a 00:05:38.913 CC module/bdev/raid/raid0.o 00:05:38.913 LIB libspdk_bdev_split.a 00:05:38.913 SO libspdk_bdev_passthru.so.6.0 00:05:38.913 SO libspdk_bdev_split.so.6.0 00:05:38.913 SYMLINK libspdk_bdev_null.so 00:05:38.913 CC module/bdev/nvme/bdev_nvme_rpc.o 00:05:39.178 SYMLINK libspdk_bdev_passthru.so 00:05:39.178 SYMLINK libspdk_bdev_split.so 00:05:39.178 CC module/bdev/raid/raid1.o 00:05:39.178 CC module/bdev/zone_block/vbdev_zone_block_rpc.o 00:05:39.178 CC module/bdev/nvme/nvme_rpc.o 00:05:39.178 LIB libspdk_bdev_malloc.a 00:05:39.178 LIB libspdk_bdev_aio.a 00:05:39.178 SO libspdk_bdev_malloc.so.6.0 00:05:39.178 SO libspdk_bdev_aio.so.6.0 00:05:39.178 CC module/bdev/raid/concat.o 00:05:39.178 SYMLINK libspdk_bdev_malloc.so 00:05:39.178 SYMLINK libspdk_bdev_aio.so 00:05:39.178 LIB libspdk_bdev_zone_block.a 00:05:39.178 SO libspdk_bdev_zone_block.so.6.0 00:05:39.439 CC module/bdev/nvme/bdev_mdns_client.o 00:05:39.439 CC module/bdev/raid/raid5f.o 00:05:39.439 SYMLINK libspdk_bdev_zone_block.so 00:05:39.439 CC module/bdev/ftl/bdev_ftl.o 00:05:39.439 CC module/bdev/ftl/bdev_ftl_rpc.o 00:05:39.439 CC module/bdev/nvme/vbdev_opal.o 00:05:39.439 CC module/bdev/iscsi/bdev_iscsi.o 00:05:39.439 CC module/bdev/virtio/bdev_virtio_scsi.o 00:05:39.439 CC module/bdev/nvme/vbdev_opal_rpc.o 00:05:39.699 CC module/bdev/nvme/bdev_nvme_cuse_rpc.o 00:05:39.699 CC module/bdev/iscsi/bdev_iscsi_rpc.o 00:05:39.699 LIB libspdk_bdev_ftl.a 00:05:39.699 SO libspdk_bdev_ftl.so.6.0 00:05:39.699 CC module/bdev/virtio/bdev_virtio_blk.o 00:05:39.699 CC module/bdev/virtio/bdev_virtio_rpc.o 00:05:39.699 SYMLINK libspdk_bdev_ftl.so 00:05:39.699 LIB libspdk_bdev_iscsi.a 00:05:39.699 SO libspdk_bdev_iscsi.so.6.0 00:05:39.959 SYMLINK libspdk_bdev_iscsi.so 00:05:39.959 LIB libspdk_bdev_raid.a 00:05:39.959 SO libspdk_bdev_raid.so.6.0 00:05:39.959 LIB libspdk_bdev_virtio.a 00:05:39.959 SO libspdk_bdev_virtio.so.6.0 00:05:39.959 SYMLINK libspdk_bdev_raid.so 00:05:40.219 SYMLINK libspdk_bdev_virtio.so 00:05:41.160 LIB libspdk_bdev_nvme.a 00:05:41.420 SO libspdk_bdev_nvme.so.7.1 00:05:41.420 SYMLINK libspdk_bdev_nvme.so 00:05:41.991 CC module/event/subsystems/scheduler/scheduler.o 00:05:41.991 CC module/event/subsystems/fsdev/fsdev.o 00:05:41.991 CC module/event/subsystems/sock/sock.o 00:05:41.991 CC module/event/subsystems/keyring/keyring.o 00:05:41.991 CC module/event/subsystems/iobuf/iobuf.o 00:05:41.991 CC module/event/subsystems/iobuf/iobuf_rpc.o 00:05:41.991 CC module/event/subsystems/vmd/vmd.o 00:05:41.991 CC module/event/subsystems/vmd/vmd_rpc.o 00:05:41.991 CC module/event/subsystems/vhost_blk/vhost_blk.o 00:05:42.251 LIB libspdk_event_scheduler.a 00:05:42.251 LIB libspdk_event_keyring.a 00:05:42.251 LIB libspdk_event_sock.a 00:05:42.251 LIB libspdk_event_vhost_blk.a 00:05:42.251 LIB libspdk_event_vmd.a 00:05:42.251 LIB libspdk_event_fsdev.a 00:05:42.251 LIB libspdk_event_iobuf.a 00:05:42.251 SO libspdk_event_keyring.so.1.0 00:05:42.251 SO libspdk_event_sock.so.5.0 00:05:42.251 SO libspdk_event_scheduler.so.4.0 00:05:42.251 SO libspdk_event_vhost_blk.so.3.0 00:05:42.251 SO libspdk_event_fsdev.so.1.0 00:05:42.251 SO libspdk_event_vmd.so.6.0 00:05:42.251 SO libspdk_event_iobuf.so.3.0 00:05:42.251 SYMLINK libspdk_event_keyring.so 00:05:42.251 SYMLINK libspdk_event_sock.so 00:05:42.251 SYMLINK libspdk_event_scheduler.so 00:05:42.251 SYMLINK libspdk_event_fsdev.so 00:05:42.251 SYMLINK libspdk_event_vhost_blk.so 00:05:42.251 SYMLINK libspdk_event_vmd.so 00:05:42.251 SYMLINK libspdk_event_iobuf.so 00:05:42.821 CC module/event/subsystems/accel/accel.o 00:05:42.821 LIB libspdk_event_accel.a 00:05:42.821 SO libspdk_event_accel.so.6.0 00:05:42.821 SYMLINK libspdk_event_accel.so 00:05:43.392 CC module/event/subsystems/bdev/bdev.o 00:05:43.392 LIB libspdk_event_bdev.a 00:05:43.652 SO libspdk_event_bdev.so.6.0 00:05:43.652 SYMLINK libspdk_event_bdev.so 00:05:43.912 CC module/event/subsystems/nbd/nbd.o 00:05:43.912 CC module/event/subsystems/scsi/scsi.o 00:05:43.912 CC module/event/subsystems/ublk/ublk.o 00:05:43.912 CC module/event/subsystems/nvmf/nvmf_rpc.o 00:05:43.912 CC module/event/subsystems/nvmf/nvmf_tgt.o 00:05:43.912 LIB libspdk_event_nbd.a 00:05:43.912 SO libspdk_event_nbd.so.6.0 00:05:43.912 LIB libspdk_event_ublk.a 00:05:44.172 LIB libspdk_event_scsi.a 00:05:44.172 SYMLINK libspdk_event_nbd.so 00:05:44.172 SO libspdk_event_ublk.so.3.0 00:05:44.172 SO libspdk_event_scsi.so.6.0 00:05:44.172 LIB libspdk_event_nvmf.a 00:05:44.172 SYMLINK libspdk_event_scsi.so 00:05:44.172 SYMLINK libspdk_event_ublk.so 00:05:44.172 SO libspdk_event_nvmf.so.6.0 00:05:44.172 SYMLINK libspdk_event_nvmf.so 00:05:44.433 CC module/event/subsystems/iscsi/iscsi.o 00:05:44.433 CC module/event/subsystems/vhost_scsi/vhost_scsi.o 00:05:44.692 LIB libspdk_event_iscsi.a 00:05:44.692 SO libspdk_event_iscsi.so.6.0 00:05:44.692 LIB libspdk_event_vhost_scsi.a 00:05:44.692 SO libspdk_event_vhost_scsi.so.3.0 00:05:44.692 SYMLINK libspdk_event_iscsi.so 00:05:44.692 SYMLINK libspdk_event_vhost_scsi.so 00:05:44.952 SO libspdk.so.6.0 00:05:44.952 SYMLINK libspdk.so 00:05:45.211 CXX app/trace/trace.o 00:05:45.211 CC app/spdk_nvme_perf/perf.o 00:05:45.211 CC app/trace_record/trace_record.o 00:05:45.211 CC app/spdk_lspci/spdk_lspci.o 00:05:45.211 CC app/spdk_nvme_identify/identify.o 00:05:45.211 CC app/iscsi_tgt/iscsi_tgt.o 00:05:45.470 CC app/nvmf_tgt/nvmf_main.o 00:05:45.470 CC app/spdk_tgt/spdk_tgt.o 00:05:45.470 CC test/thread/poller_perf/poller_perf.o 00:05:45.470 CC examples/util/zipf/zipf.o 00:05:45.470 LINK spdk_lspci 00:05:45.470 LINK nvmf_tgt 00:05:45.470 LINK poller_perf 00:05:45.470 LINK iscsi_tgt 00:05:45.470 LINK spdk_tgt 00:05:45.470 LINK zipf 00:05:45.470 LINK spdk_trace_record 00:05:45.734 CC app/spdk_nvme_discover/discovery_aer.o 00:05:45.734 LINK spdk_trace 00:05:45.734 CC app/spdk_top/spdk_top.o 00:05:45.734 CC app/spdk_dd/spdk_dd.o 00:05:45.734 LINK spdk_nvme_discover 00:05:45.734 CC examples/ioat/perf/perf.o 00:05:45.734 CC test/dma/test_dma/test_dma.o 00:05:45.995 CC app/fio/nvme/fio_plugin.o 00:05:45.995 CC examples/vmd/lsvmd/lsvmd.o 00:05:45.995 CC examples/ioat/verify/verify.o 00:05:45.995 LINK ioat_perf 00:05:45.995 LINK lsvmd 00:05:46.255 LINK spdk_dd 00:05:46.255 CC test/app/bdev_svc/bdev_svc.o 00:05:46.255 LINK verify 00:05:46.255 LINK spdk_nvme_perf 00:05:46.255 LINK spdk_nvme_identify 00:05:46.255 CC examples/vmd/led/led.o 00:05:46.255 LINK test_dma 00:05:46.255 LINK bdev_svc 00:05:46.255 CC test/app/fuzz/nvme_fuzz/nvme_fuzz.o 00:05:46.255 CC app/fio/bdev/fio_plugin.o 00:05:46.515 CC test/app/fuzz/iscsi_fuzz/iscsi_fuzz.o 00:05:46.515 CC test/app/fuzz/vhost_fuzz/vhost_fuzz_rpc.o 00:05:46.515 LINK led 00:05:46.515 LINK spdk_nvme 00:05:46.515 TEST_HEADER include/spdk/accel.h 00:05:46.515 TEST_HEADER include/spdk/accel_module.h 00:05:46.515 TEST_HEADER include/spdk/assert.h 00:05:46.515 TEST_HEADER include/spdk/barrier.h 00:05:46.515 TEST_HEADER include/spdk/base64.h 00:05:46.515 TEST_HEADER include/spdk/bdev.h 00:05:46.515 TEST_HEADER include/spdk/bdev_module.h 00:05:46.515 TEST_HEADER include/spdk/bdev_zone.h 00:05:46.515 TEST_HEADER include/spdk/bit_array.h 00:05:46.515 TEST_HEADER include/spdk/bit_pool.h 00:05:46.515 TEST_HEADER include/spdk/blob_bdev.h 00:05:46.515 TEST_HEADER include/spdk/blobfs_bdev.h 00:05:46.515 TEST_HEADER include/spdk/blobfs.h 00:05:46.515 TEST_HEADER include/spdk/blob.h 00:05:46.515 TEST_HEADER include/spdk/conf.h 00:05:46.515 TEST_HEADER include/spdk/config.h 00:05:46.515 TEST_HEADER include/spdk/cpuset.h 00:05:46.515 TEST_HEADER include/spdk/crc16.h 00:05:46.515 TEST_HEADER include/spdk/crc32.h 00:05:46.515 TEST_HEADER include/spdk/crc64.h 00:05:46.515 TEST_HEADER include/spdk/dif.h 00:05:46.515 TEST_HEADER include/spdk/dma.h 00:05:46.515 TEST_HEADER include/spdk/endian.h 00:05:46.515 TEST_HEADER include/spdk/env_dpdk.h 00:05:46.515 TEST_HEADER include/spdk/env.h 00:05:46.515 CC test/app/histogram_perf/histogram_perf.o 00:05:46.515 TEST_HEADER include/spdk/event.h 00:05:46.515 TEST_HEADER include/spdk/fd_group.h 00:05:46.515 TEST_HEADER include/spdk/fd.h 00:05:46.515 TEST_HEADER include/spdk/file.h 00:05:46.515 CC test/app/fuzz/vhost_fuzz/vhost_fuzz.o 00:05:46.515 TEST_HEADER include/spdk/fsdev.h 00:05:46.515 TEST_HEADER include/spdk/fsdev_module.h 00:05:46.515 TEST_HEADER include/spdk/ftl.h 00:05:46.515 TEST_HEADER include/spdk/fuse_dispatcher.h 00:05:46.515 TEST_HEADER include/spdk/gpt_spec.h 00:05:46.515 TEST_HEADER include/spdk/hexlify.h 00:05:46.515 TEST_HEADER include/spdk/histogram_data.h 00:05:46.515 TEST_HEADER include/spdk/idxd.h 00:05:46.515 TEST_HEADER include/spdk/idxd_spec.h 00:05:46.515 TEST_HEADER include/spdk/init.h 00:05:46.515 TEST_HEADER include/spdk/ioat.h 00:05:46.515 TEST_HEADER include/spdk/ioat_spec.h 00:05:46.515 TEST_HEADER include/spdk/iscsi_spec.h 00:05:46.515 TEST_HEADER include/spdk/json.h 00:05:46.515 TEST_HEADER include/spdk/jsonrpc.h 00:05:46.515 TEST_HEADER include/spdk/keyring.h 00:05:46.515 TEST_HEADER include/spdk/keyring_module.h 00:05:46.515 TEST_HEADER include/spdk/likely.h 00:05:46.515 TEST_HEADER include/spdk/log.h 00:05:46.515 TEST_HEADER include/spdk/lvol.h 00:05:46.515 TEST_HEADER include/spdk/md5.h 00:05:46.515 TEST_HEADER include/spdk/memory.h 00:05:46.515 TEST_HEADER include/spdk/mmio.h 00:05:46.515 TEST_HEADER include/spdk/nbd.h 00:05:46.515 TEST_HEADER include/spdk/net.h 00:05:46.515 TEST_HEADER include/spdk/notify.h 00:05:46.515 TEST_HEADER include/spdk/nvme.h 00:05:46.515 TEST_HEADER include/spdk/nvme_intel.h 00:05:46.515 CC app/vhost/vhost.o 00:05:46.515 TEST_HEADER include/spdk/nvme_ocssd.h 00:05:46.515 TEST_HEADER include/spdk/nvme_ocssd_spec.h 00:05:46.515 TEST_HEADER include/spdk/nvme_spec.h 00:05:46.515 TEST_HEADER include/spdk/nvme_zns.h 00:05:46.515 TEST_HEADER include/spdk/nvmf_cmd.h 00:05:46.515 TEST_HEADER include/spdk/nvmf_fc_spec.h 00:05:46.515 TEST_HEADER include/spdk/nvmf.h 00:05:46.515 TEST_HEADER include/spdk/nvmf_spec.h 00:05:46.515 TEST_HEADER include/spdk/nvmf_transport.h 00:05:46.515 TEST_HEADER include/spdk/opal.h 00:05:46.515 TEST_HEADER include/spdk/opal_spec.h 00:05:46.515 TEST_HEADER include/spdk/pci_ids.h 00:05:46.515 TEST_HEADER include/spdk/pipe.h 00:05:46.515 TEST_HEADER include/spdk/queue.h 00:05:46.515 TEST_HEADER include/spdk/reduce.h 00:05:46.515 TEST_HEADER include/spdk/rpc.h 00:05:46.515 TEST_HEADER include/spdk/scheduler.h 00:05:46.515 TEST_HEADER include/spdk/scsi.h 00:05:46.515 TEST_HEADER include/spdk/scsi_spec.h 00:05:46.515 TEST_HEADER include/spdk/sock.h 00:05:46.515 TEST_HEADER include/spdk/stdinc.h 00:05:46.774 TEST_HEADER include/spdk/string.h 00:05:46.775 TEST_HEADER include/spdk/thread.h 00:05:46.775 TEST_HEADER include/spdk/trace.h 00:05:46.775 TEST_HEADER include/spdk/trace_parser.h 00:05:46.775 TEST_HEADER include/spdk/tree.h 00:05:46.775 TEST_HEADER include/spdk/ublk.h 00:05:46.775 TEST_HEADER include/spdk/util.h 00:05:46.775 TEST_HEADER include/spdk/uuid.h 00:05:46.775 TEST_HEADER include/spdk/version.h 00:05:46.775 TEST_HEADER include/spdk/vfio_user_pci.h 00:05:46.775 TEST_HEADER include/spdk/vfio_user_spec.h 00:05:46.775 TEST_HEADER include/spdk/vhost.h 00:05:46.775 TEST_HEADER include/spdk/vmd.h 00:05:46.775 TEST_HEADER include/spdk/xor.h 00:05:46.775 TEST_HEADER include/spdk/zipf.h 00:05:46.775 CXX test/cpp_headers/accel.o 00:05:46.775 LINK histogram_perf 00:05:46.775 LINK spdk_top 00:05:46.775 CC examples/interrupt_tgt/interrupt_tgt.o 00:05:46.775 LINK nvme_fuzz 00:05:46.775 CC examples/idxd/perf/perf.o 00:05:46.775 LINK vhost 00:05:46.775 CXX test/cpp_headers/accel_module.o 00:05:46.775 CXX test/cpp_headers/assert.o 00:05:46.775 LINK spdk_bdev 00:05:46.775 LINK interrupt_tgt 00:05:47.034 CC test/app/jsoncat/jsoncat.o 00:05:47.034 CXX test/cpp_headers/barrier.o 00:05:47.034 CC test/app/stub/stub.o 00:05:47.034 LINK vhost_fuzz 00:05:47.034 LINK jsoncat 00:05:47.034 LINK idxd_perf 00:05:47.034 CXX test/cpp_headers/base64.o 00:05:47.034 CXX test/cpp_headers/bdev.o 00:05:47.034 LINK stub 00:05:47.294 CC test/event/event_perf/event_perf.o 00:05:47.294 CC examples/thread/thread/thread_ex.o 00:05:47.294 CC test/nvme/aer/aer.o 00:05:47.294 CC test/env/mem_callbacks/mem_callbacks.o 00:05:47.294 CC test/nvme/reset/reset.o 00:05:47.294 CXX test/cpp_headers/bdev_module.o 00:05:47.294 CC test/nvme/sgl/sgl.o 00:05:47.294 LINK event_perf 00:05:47.294 CC test/nvme/e2edp/nvme_dp.o 00:05:47.294 CC test/nvme/overhead/overhead.o 00:05:47.553 LINK mem_callbacks 00:05:47.554 LINK thread 00:05:47.554 LINK aer 00:05:47.554 CXX test/cpp_headers/bdev_zone.o 00:05:47.554 CC test/event/reactor/reactor.o 00:05:47.554 LINK reset 00:05:47.554 LINK sgl 00:05:47.554 LINK nvme_dp 00:05:47.554 CC test/env/vtophys/vtophys.o 00:05:47.554 LINK reactor 00:05:47.554 LINK overhead 00:05:47.554 CXX test/cpp_headers/bit_array.o 00:05:47.813 CC test/env/env_dpdk_post_init/env_dpdk_post_init.o 00:05:47.813 LINK vtophys 00:05:47.813 CC examples/sock/hello_world/hello_sock.o 00:05:47.813 CC test/nvme/err_injection/err_injection.o 00:05:47.813 CXX test/cpp_headers/bit_pool.o 00:05:47.813 CC test/event/reactor_perf/reactor_perf.o 00:05:47.813 LINK env_dpdk_post_init 00:05:47.813 CC test/rpc_client/rpc_client_test.o 00:05:47.813 CC test/event/app_repeat/app_repeat.o 00:05:47.813 CC test/nvme/startup/startup.o 00:05:48.072 CXX test/cpp_headers/blob_bdev.o 00:05:48.072 LINK reactor_perf 00:05:48.072 LINK err_injection 00:05:48.072 LINK hello_sock 00:05:48.072 CC test/event/scheduler/scheduler.o 00:05:48.072 LINK app_repeat 00:05:48.072 LINK rpc_client_test 00:05:48.072 LINK startup 00:05:48.072 CC test/env/memory/memory_ut.o 00:05:48.072 CXX test/cpp_headers/blobfs_bdev.o 00:05:48.072 CXX test/cpp_headers/blobfs.o 00:05:48.072 CXX test/cpp_headers/blob.o 00:05:48.332 LINK iscsi_fuzz 00:05:48.332 CXX test/cpp_headers/conf.o 00:05:48.332 LINK scheduler 00:05:48.332 CC examples/fsdev/hello_world/hello_fsdev.o 00:05:48.332 CC test/nvme/reserve/reserve.o 00:05:48.332 CXX test/cpp_headers/config.o 00:05:48.332 CXX test/cpp_headers/cpuset.o 00:05:48.332 CC test/accel/dif/dif.o 00:05:48.332 CC test/env/pci/pci_ut.o 00:05:48.332 CC examples/accel/perf/accel_perf.o 00:05:48.591 LINK reserve 00:05:48.591 LINK hello_fsdev 00:05:48.591 CC test/nvme/simple_copy/simple_copy.o 00:05:48.591 CXX test/cpp_headers/crc16.o 00:05:48.591 CC examples/blob/hello_world/hello_blob.o 00:05:48.591 CC examples/nvme/hello_world/hello_world.o 00:05:48.591 CXX test/cpp_headers/crc32.o 00:05:48.851 CC examples/nvme/reconnect/reconnect.o 00:05:48.851 LINK simple_copy 00:05:48.851 LINK hello_blob 00:05:48.851 CC test/nvme/connect_stress/connect_stress.o 00:05:48.851 LINK hello_world 00:05:48.851 LINK memory_ut 00:05:48.851 LINK pci_ut 00:05:48.851 CXX test/cpp_headers/crc64.o 00:05:48.851 LINK accel_perf 00:05:48.851 LINK connect_stress 00:05:49.110 CC test/nvme/boot_partition/boot_partition.o 00:05:49.110 CXX test/cpp_headers/dif.o 00:05:49.110 CC examples/blob/cli/blobcli.o 00:05:49.110 CC test/nvme/compliance/nvme_compliance.o 00:05:49.110 LINK dif 00:05:49.110 LINK reconnect 00:05:49.110 CC test/nvme/fused_ordering/fused_ordering.o 00:05:49.110 LINK boot_partition 00:05:49.110 CXX test/cpp_headers/dma.o 00:05:49.110 CC test/nvme/doorbell_aers/doorbell_aers.o 00:05:49.110 CC examples/nvme/nvme_manage/nvme_manage.o 00:05:49.110 CC test/blobfs/mkfs/mkfs.o 00:05:49.369 CXX test/cpp_headers/endian.o 00:05:49.369 LINK fused_ordering 00:05:49.369 LINK doorbell_aers 00:05:49.369 LINK nvme_compliance 00:05:49.369 CC examples/nvme/hotplug/hotplug.o 00:05:49.369 CC examples/nvme/arbitration/arbitration.o 00:05:49.369 LINK mkfs 00:05:49.369 CXX test/cpp_headers/env_dpdk.o 00:05:49.369 CC test/lvol/esnap/esnap.o 00:05:49.630 LINK blobcli 00:05:49.630 CXX test/cpp_headers/env.o 00:05:49.630 CC examples/nvme/cmb_copy/cmb_copy.o 00:05:49.630 LINK hotplug 00:05:49.630 CC test/nvme/fdp/fdp.o 00:05:49.630 LINK nvme_manage 00:05:49.630 CC test/bdev/bdevio/bdevio.o 00:05:49.630 CC test/nvme/cuse/cuse.o 00:05:49.630 CXX test/cpp_headers/event.o 00:05:49.630 LINK arbitration 00:05:49.890 LINK cmb_copy 00:05:49.890 CC examples/nvme/abort/abort.o 00:05:49.890 CXX test/cpp_headers/fd_group.o 00:05:49.890 CC examples/bdev/hello_world/hello_bdev.o 00:05:49.890 CXX test/cpp_headers/fd.o 00:05:49.890 CC examples/bdev/bdevperf/bdevperf.o 00:05:49.890 LINK fdp 00:05:49.890 CXX test/cpp_headers/file.o 00:05:50.150 CC examples/nvme/pmr_persistence/pmr_persistence.o 00:05:50.150 CXX test/cpp_headers/fsdev.o 00:05:50.150 LINK bdevio 00:05:50.150 CXX test/cpp_headers/fsdev_module.o 00:05:50.150 LINK hello_bdev 00:05:50.150 CXX test/cpp_headers/ftl.o 00:05:50.150 LINK pmr_persistence 00:05:50.150 LINK abort 00:05:50.150 CXX test/cpp_headers/fuse_dispatcher.o 00:05:50.150 CXX test/cpp_headers/gpt_spec.o 00:05:50.410 CXX test/cpp_headers/hexlify.o 00:05:50.410 CXX test/cpp_headers/histogram_data.o 00:05:50.410 CXX test/cpp_headers/idxd.o 00:05:50.410 CXX test/cpp_headers/idxd_spec.o 00:05:50.410 CXX test/cpp_headers/init.o 00:05:50.410 CXX test/cpp_headers/ioat.o 00:05:50.410 CXX test/cpp_headers/ioat_spec.o 00:05:50.410 CXX test/cpp_headers/iscsi_spec.o 00:05:50.410 CXX test/cpp_headers/json.o 00:05:50.410 CXX test/cpp_headers/jsonrpc.o 00:05:50.410 CXX test/cpp_headers/keyring.o 00:05:50.410 CXX test/cpp_headers/keyring_module.o 00:05:50.410 CXX test/cpp_headers/likely.o 00:05:50.410 CXX test/cpp_headers/log.o 00:05:50.671 CXX test/cpp_headers/lvol.o 00:05:50.671 CXX test/cpp_headers/md5.o 00:05:50.671 CXX test/cpp_headers/memory.o 00:05:50.671 CXX test/cpp_headers/mmio.o 00:05:50.671 CXX test/cpp_headers/nbd.o 00:05:50.671 CXX test/cpp_headers/net.o 00:05:50.671 CXX test/cpp_headers/notify.o 00:05:50.671 CXX test/cpp_headers/nvme.o 00:05:50.671 CXX test/cpp_headers/nvme_intel.o 00:05:50.671 CXX test/cpp_headers/nvme_ocssd.o 00:05:50.671 LINK bdevperf 00:05:50.671 CXX test/cpp_headers/nvme_ocssd_spec.o 00:05:50.671 CXX test/cpp_headers/nvme_spec.o 00:05:50.931 CXX test/cpp_headers/nvme_zns.o 00:05:50.931 CXX test/cpp_headers/nvmf_cmd.o 00:05:50.931 CXX test/cpp_headers/nvmf_fc_spec.o 00:05:50.931 CXX test/cpp_headers/nvmf.o 00:05:50.931 CXX test/cpp_headers/nvmf_spec.o 00:05:50.931 CXX test/cpp_headers/nvmf_transport.o 00:05:50.931 CXX test/cpp_headers/opal.o 00:05:50.931 LINK cuse 00:05:50.931 CXX test/cpp_headers/opal_spec.o 00:05:50.931 CXX test/cpp_headers/pci_ids.o 00:05:50.931 CXX test/cpp_headers/pipe.o 00:05:50.931 CXX test/cpp_headers/queue.o 00:05:50.931 CXX test/cpp_headers/reduce.o 00:05:51.191 CXX test/cpp_headers/rpc.o 00:05:51.191 CXX test/cpp_headers/scheduler.o 00:05:51.191 CXX test/cpp_headers/scsi.o 00:05:51.191 CXX test/cpp_headers/scsi_spec.o 00:05:51.191 CXX test/cpp_headers/sock.o 00:05:51.191 CXX test/cpp_headers/stdinc.o 00:05:51.191 CC examples/nvmf/nvmf/nvmf.o 00:05:51.191 CXX test/cpp_headers/string.o 00:05:51.191 CXX test/cpp_headers/thread.o 00:05:51.191 CXX test/cpp_headers/trace.o 00:05:51.191 CXX test/cpp_headers/trace_parser.o 00:05:51.191 CXX test/cpp_headers/tree.o 00:05:51.191 CXX test/cpp_headers/ublk.o 00:05:51.191 CXX test/cpp_headers/util.o 00:05:51.191 CXX test/cpp_headers/uuid.o 00:05:51.191 CXX test/cpp_headers/version.o 00:05:51.451 CXX test/cpp_headers/vfio_user_pci.o 00:05:51.451 CXX test/cpp_headers/vfio_user_spec.o 00:05:51.451 CXX test/cpp_headers/vhost.o 00:05:51.451 CXX test/cpp_headers/vmd.o 00:05:51.451 CXX test/cpp_headers/xor.o 00:05:51.451 CXX test/cpp_headers/zipf.o 00:05:51.451 LINK nvmf 00:05:54.747 LINK esnap 00:05:55.317 00:05:55.317 real 1m22.143s 00:05:55.317 user 6m3.325s 00:05:55.317 sys 1m14.762s 00:05:55.317 15:11:45 make -- common/autotest_common.sh@1130 -- $ xtrace_disable 00:05:55.317 15:11:45 make -- common/autotest_common.sh@10 -- $ set +x 00:05:55.317 ************************************ 00:05:55.317 END TEST make 00:05:55.317 ************************************ 00:05:55.317 15:11:45 -- spdk/autobuild.sh@1 -- $ stop_monitor_resources 00:05:55.317 15:11:45 -- pm/common@29 -- $ signal_monitor_resources TERM 00:05:55.317 15:11:45 -- pm/common@40 -- $ local monitor pid pids signal=TERM 00:05:55.317 15:11:45 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:05:55.317 15:11:45 -- pm/common@43 -- $ [[ -e /home/vagrant/spdk_repo/spdk/../output/power/collect-cpu-load.pid ]] 00:05:55.317 15:11:45 -- pm/common@44 -- $ pid=6209 00:05:55.317 15:11:45 -- pm/common@50 -- $ kill -TERM 6209 00:05:55.317 15:11:45 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:05:55.317 15:11:45 -- pm/common@43 -- $ [[ -e /home/vagrant/spdk_repo/spdk/../output/power/collect-vmstat.pid ]] 00:05:55.317 15:11:45 -- pm/common@44 -- $ pid=6210 00:05:55.317 15:11:45 -- pm/common@50 -- $ kill -TERM 6210 00:05:55.317 15:11:45 -- spdk/autorun.sh@26 -- $ (( SPDK_TEST_UNITTEST == 1 || SPDK_RUN_FUNCTIONAL_TEST == 1 )) 00:05:55.317 15:11:45 -- spdk/autorun.sh@27 -- $ sudo -E /home/vagrant/spdk_repo/spdk/autotest.sh /home/vagrant/spdk_repo/autorun-spdk.conf 00:05:55.317 15:11:45 -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:05:55.317 15:11:45 -- common/autotest_common.sh@1693 -- # lcov --version 00:05:55.317 15:11:45 -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:05:55.577 15:11:45 -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:05:55.577 15:11:45 -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:55.577 15:11:45 -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:55.577 15:11:45 -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:55.577 15:11:45 -- scripts/common.sh@336 -- # IFS=.-: 00:05:55.577 15:11:45 -- scripts/common.sh@336 -- # read -ra ver1 00:05:55.577 15:11:45 -- scripts/common.sh@337 -- # IFS=.-: 00:05:55.577 15:11:45 -- scripts/common.sh@337 -- # read -ra ver2 00:05:55.577 15:11:45 -- scripts/common.sh@338 -- # local 'op=<' 00:05:55.577 15:11:45 -- scripts/common.sh@340 -- # ver1_l=2 00:05:55.577 15:11:45 -- scripts/common.sh@341 -- # ver2_l=1 00:05:55.577 15:11:45 -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:55.577 15:11:45 -- scripts/common.sh@344 -- # case "$op" in 00:05:55.577 15:11:45 -- scripts/common.sh@345 -- # : 1 00:05:55.577 15:11:45 -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:55.577 15:11:45 -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:55.577 15:11:45 -- scripts/common.sh@365 -- # decimal 1 00:05:55.577 15:11:45 -- scripts/common.sh@353 -- # local d=1 00:05:55.577 15:11:45 -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:55.577 15:11:45 -- scripts/common.sh@355 -- # echo 1 00:05:55.577 15:11:45 -- scripts/common.sh@365 -- # ver1[v]=1 00:05:55.577 15:11:45 -- scripts/common.sh@366 -- # decimal 2 00:05:55.577 15:11:45 -- scripts/common.sh@353 -- # local d=2 00:05:55.577 15:11:45 -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:55.577 15:11:45 -- scripts/common.sh@355 -- # echo 2 00:05:55.577 15:11:45 -- scripts/common.sh@366 -- # ver2[v]=2 00:05:55.577 15:11:45 -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:55.577 15:11:45 -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:55.577 15:11:45 -- scripts/common.sh@368 -- # return 0 00:05:55.577 15:11:45 -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:55.577 15:11:45 -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:05:55.577 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:55.577 --rc genhtml_branch_coverage=1 00:05:55.577 --rc genhtml_function_coverage=1 00:05:55.577 --rc genhtml_legend=1 00:05:55.577 --rc geninfo_all_blocks=1 00:05:55.577 --rc geninfo_unexecuted_blocks=1 00:05:55.577 00:05:55.577 ' 00:05:55.577 15:11:45 -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:05:55.577 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:55.577 --rc genhtml_branch_coverage=1 00:05:55.577 --rc genhtml_function_coverage=1 00:05:55.577 --rc genhtml_legend=1 00:05:55.577 --rc geninfo_all_blocks=1 00:05:55.577 --rc geninfo_unexecuted_blocks=1 00:05:55.577 00:05:55.577 ' 00:05:55.577 15:11:45 -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:05:55.577 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:55.577 --rc genhtml_branch_coverage=1 00:05:55.577 --rc genhtml_function_coverage=1 00:05:55.577 --rc genhtml_legend=1 00:05:55.577 --rc geninfo_all_blocks=1 00:05:55.577 --rc geninfo_unexecuted_blocks=1 00:05:55.577 00:05:55.577 ' 00:05:55.577 15:11:45 -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:05:55.577 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:55.577 --rc genhtml_branch_coverage=1 00:05:55.577 --rc genhtml_function_coverage=1 00:05:55.577 --rc genhtml_legend=1 00:05:55.577 --rc geninfo_all_blocks=1 00:05:55.577 --rc geninfo_unexecuted_blocks=1 00:05:55.577 00:05:55.577 ' 00:05:55.577 15:11:45 -- spdk/autotest.sh@25 -- # source /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh 00:05:55.577 15:11:45 -- nvmf/common.sh@7 -- # uname -s 00:05:55.577 15:11:45 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:05:55.577 15:11:45 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:05:55.577 15:11:45 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:05:55.577 15:11:45 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:05:55.577 15:11:45 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:05:55.577 15:11:45 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:05:55.577 15:11:45 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:05:55.577 15:11:45 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:05:55.577 15:11:45 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:05:55.577 15:11:45 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:05:55.577 15:11:45 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:05233599-7658-47bb-b32b-d880580825e2 00:05:55.577 15:11:45 -- nvmf/common.sh@18 -- # NVME_HOSTID=05233599-7658-47bb-b32b-d880580825e2 00:05:55.577 15:11:45 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:05:55.577 15:11:45 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:05:55.577 15:11:45 -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:05:55.577 15:11:45 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:05:55.577 15:11:45 -- nvmf/common.sh@49 -- # source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:05:55.577 15:11:45 -- scripts/common.sh@15 -- # shopt -s extglob 00:05:55.577 15:11:45 -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:05:55.577 15:11:45 -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:05:55.577 15:11:45 -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:05:55.577 15:11:45 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:55.577 15:11:45 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:55.577 15:11:45 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:55.577 15:11:45 -- paths/export.sh@5 -- # export PATH 00:05:55.577 15:11:45 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:55.577 15:11:45 -- nvmf/common.sh@51 -- # : 0 00:05:55.577 15:11:45 -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:05:55.577 15:11:45 -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:05:55.577 15:11:45 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:05:55.577 15:11:45 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:05:55.577 15:11:45 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:05:55.577 15:11:45 -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:05:55.577 /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:05:55.577 15:11:45 -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:05:55.577 15:11:45 -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:05:55.577 15:11:45 -- nvmf/common.sh@55 -- # have_pci_nics=0 00:05:55.577 15:11:45 -- spdk/autotest.sh@27 -- # '[' 0 -ne 0 ']' 00:05:55.577 15:11:45 -- spdk/autotest.sh@32 -- # uname -s 00:05:55.577 15:11:45 -- spdk/autotest.sh@32 -- # '[' Linux = Linux ']' 00:05:55.577 15:11:45 -- spdk/autotest.sh@33 -- # old_core_pattern='|/usr/lib/systemd/systemd-coredump %P %u %g %s %t %c %h' 00:05:55.577 15:11:45 -- spdk/autotest.sh@34 -- # mkdir -p /home/vagrant/spdk_repo/spdk/../output/coredumps 00:05:55.577 15:11:45 -- spdk/autotest.sh@39 -- # echo '|/home/vagrant/spdk_repo/spdk/scripts/core-collector.sh %P %s %t' 00:05:55.577 15:11:45 -- spdk/autotest.sh@40 -- # echo /home/vagrant/spdk_repo/spdk/../output/coredumps 00:05:55.577 15:11:45 -- spdk/autotest.sh@44 -- # modprobe nbd 00:05:55.577 15:11:45 -- spdk/autotest.sh@46 -- # type -P udevadm 00:05:55.577 15:11:45 -- spdk/autotest.sh@46 -- # udevadm=/usr/sbin/udevadm 00:05:55.577 15:11:45 -- spdk/autotest.sh@48 -- # udevadm_pid=66627 00:05:55.577 15:11:45 -- spdk/autotest.sh@47 -- # /usr/sbin/udevadm monitor --property 00:05:55.577 15:11:45 -- spdk/autotest.sh@53 -- # start_monitor_resources 00:05:55.577 15:11:45 -- pm/common@17 -- # local monitor 00:05:55.577 15:11:45 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:05:55.577 15:11:45 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:05:55.577 15:11:45 -- pm/common@21 -- # date +%s 00:05:55.577 15:11:45 -- pm/common@25 -- # sleep 1 00:05:55.577 15:11:45 -- pm/common@21 -- # date +%s 00:05:55.577 15:11:45 -- pm/common@21 -- # /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-cpu-load -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autotest.sh.1732029105 00:05:55.577 15:11:45 -- pm/common@21 -- # /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-vmstat -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autotest.sh.1732029105 00:05:55.577 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autotest.sh.1732029105_collect-cpu-load.pm.log 00:05:55.577 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autotest.sh.1732029105_collect-vmstat.pm.log 00:05:56.964 15:11:46 -- spdk/autotest.sh@55 -- # trap 'autotest_cleanup || :; exit 1' SIGINT SIGTERM EXIT 00:05:56.964 15:11:46 -- spdk/autotest.sh@57 -- # timing_enter autotest 00:05:56.964 15:11:46 -- common/autotest_common.sh@726 -- # xtrace_disable 00:05:56.964 15:11:46 -- common/autotest_common.sh@10 -- # set +x 00:05:56.964 15:11:46 -- spdk/autotest.sh@59 -- # create_test_list 00:05:56.964 15:11:46 -- common/autotest_common.sh@752 -- # xtrace_disable 00:05:56.964 15:11:46 -- common/autotest_common.sh@10 -- # set +x 00:05:56.964 15:11:46 -- spdk/autotest.sh@61 -- # dirname /home/vagrant/spdk_repo/spdk/autotest.sh 00:05:56.964 15:11:46 -- spdk/autotest.sh@61 -- # readlink -f /home/vagrant/spdk_repo/spdk 00:05:56.964 15:11:46 -- spdk/autotest.sh@61 -- # src=/home/vagrant/spdk_repo/spdk 00:05:56.964 15:11:46 -- spdk/autotest.sh@62 -- # out=/home/vagrant/spdk_repo/spdk/../output 00:05:56.964 15:11:46 -- spdk/autotest.sh@63 -- # cd /home/vagrant/spdk_repo/spdk 00:05:56.964 15:11:46 -- spdk/autotest.sh@65 -- # freebsd_update_contigmem_mod 00:05:56.964 15:11:46 -- common/autotest_common.sh@1457 -- # uname 00:05:56.964 15:11:46 -- common/autotest_common.sh@1457 -- # '[' Linux = FreeBSD ']' 00:05:56.964 15:11:46 -- spdk/autotest.sh@66 -- # freebsd_set_maxsock_buf 00:05:56.964 15:11:46 -- common/autotest_common.sh@1477 -- # uname 00:05:56.964 15:11:46 -- common/autotest_common.sh@1477 -- # [[ Linux = FreeBSD ]] 00:05:56.964 15:11:46 -- spdk/autotest.sh@68 -- # [[ y == y ]] 00:05:56.964 15:11:46 -- spdk/autotest.sh@70 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 --version 00:05:56.964 lcov: LCOV version 1.15 00:05:56.964 15:11:47 -- spdk/autotest.sh@72 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -c --no-external -i -t Baseline -d /home/vagrant/spdk_repo/spdk -o /home/vagrant/spdk_repo/spdk/../output/cov_base.info 00:06:11.865 /home/vagrant/spdk_repo/spdk/lib/nvme/nvme_stubs.gcno:no functions found 00:06:11.865 geninfo: WARNING: GCOV did not produce any data for /home/vagrant/spdk_repo/spdk/lib/nvme/nvme_stubs.gcno 00:06:26.791 15:12:15 -- spdk/autotest.sh@76 -- # timing_enter pre_cleanup 00:06:26.791 15:12:15 -- common/autotest_common.sh@726 -- # xtrace_disable 00:06:26.791 15:12:15 -- common/autotest_common.sh@10 -- # set +x 00:06:26.791 15:12:15 -- spdk/autotest.sh@78 -- # rm -f 00:06:26.791 15:12:15 -- spdk/autotest.sh@81 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh reset 00:06:26.791 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:06:26.791 0000:00:11.0 (1b36 0010): Already using the nvme driver 00:06:26.791 0000:00:10.0 (1b36 0010): Already using the nvme driver 00:06:26.791 15:12:15 -- spdk/autotest.sh@83 -- # get_zoned_devs 00:06:26.791 15:12:15 -- common/autotest_common.sh@1657 -- # zoned_devs=() 00:06:26.792 15:12:15 -- common/autotest_common.sh@1657 -- # local -gA zoned_devs 00:06:26.792 15:12:15 -- common/autotest_common.sh@1658 -- # local nvme bdf 00:06:26.792 15:12:15 -- common/autotest_common.sh@1660 -- # for nvme in /sys/block/nvme* 00:06:26.792 15:12:15 -- common/autotest_common.sh@1661 -- # is_block_zoned nvme0n1 00:06:26.792 15:12:15 -- common/autotest_common.sh@1650 -- # local device=nvme0n1 00:06:26.792 15:12:15 -- common/autotest_common.sh@1652 -- # [[ -e /sys/block/nvme0n1/queue/zoned ]] 00:06:26.792 15:12:15 -- common/autotest_common.sh@1653 -- # [[ none != none ]] 00:06:26.792 15:12:15 -- common/autotest_common.sh@1660 -- # for nvme in /sys/block/nvme* 00:06:26.792 15:12:15 -- common/autotest_common.sh@1661 -- # is_block_zoned nvme1n1 00:06:26.792 15:12:15 -- common/autotest_common.sh@1650 -- # local device=nvme1n1 00:06:26.792 15:12:15 -- common/autotest_common.sh@1652 -- # [[ -e /sys/block/nvme1n1/queue/zoned ]] 00:06:26.792 15:12:15 -- common/autotest_common.sh@1653 -- # [[ none != none ]] 00:06:26.792 15:12:15 -- common/autotest_common.sh@1660 -- # for nvme in /sys/block/nvme* 00:06:26.792 15:12:15 -- common/autotest_common.sh@1661 -- # is_block_zoned nvme1n2 00:06:26.792 15:12:15 -- common/autotest_common.sh@1650 -- # local device=nvme1n2 00:06:26.792 15:12:15 -- common/autotest_common.sh@1652 -- # [[ -e /sys/block/nvme1n2/queue/zoned ]] 00:06:26.792 15:12:15 -- common/autotest_common.sh@1653 -- # [[ none != none ]] 00:06:26.792 15:12:15 -- common/autotest_common.sh@1660 -- # for nvme in /sys/block/nvme* 00:06:26.792 15:12:15 -- common/autotest_common.sh@1661 -- # is_block_zoned nvme1n3 00:06:26.792 15:12:15 -- common/autotest_common.sh@1650 -- # local device=nvme1n3 00:06:26.792 15:12:15 -- common/autotest_common.sh@1652 -- # [[ -e /sys/block/nvme1n3/queue/zoned ]] 00:06:26.792 15:12:15 -- common/autotest_common.sh@1653 -- # [[ none != none ]] 00:06:26.792 15:12:15 -- spdk/autotest.sh@85 -- # (( 0 > 0 )) 00:06:26.792 15:12:15 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:06:26.792 15:12:15 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:06:26.792 15:12:15 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme0n1 00:06:26.792 15:12:15 -- scripts/common.sh@381 -- # local block=/dev/nvme0n1 pt 00:06:26.792 15:12:15 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme0n1 00:06:26.792 No valid GPT data, bailing 00:06:26.792 15:12:16 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme0n1 00:06:26.792 15:12:16 -- scripts/common.sh@394 -- # pt= 00:06:26.792 15:12:16 -- scripts/common.sh@395 -- # return 1 00:06:26.792 15:12:16 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme0n1 bs=1M count=1 00:06:26.792 1+0 records in 00:06:26.792 1+0 records out 00:06:26.792 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00702467 s, 149 MB/s 00:06:26.792 15:12:16 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:06:26.792 15:12:16 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:06:26.792 15:12:16 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme1n1 00:06:26.792 15:12:16 -- scripts/common.sh@381 -- # local block=/dev/nvme1n1 pt 00:06:26.792 15:12:16 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme1n1 00:06:26.792 No valid GPT data, bailing 00:06:26.792 15:12:16 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme1n1 00:06:26.792 15:12:16 -- scripts/common.sh@394 -- # pt= 00:06:26.792 15:12:16 -- scripts/common.sh@395 -- # return 1 00:06:26.792 15:12:16 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme1n1 bs=1M count=1 00:06:26.792 1+0 records in 00:06:26.792 1+0 records out 00:06:26.792 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0045399 s, 231 MB/s 00:06:26.792 15:12:16 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:06:26.792 15:12:16 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:06:26.792 15:12:16 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme1n2 00:06:26.792 15:12:16 -- scripts/common.sh@381 -- # local block=/dev/nvme1n2 pt 00:06:26.792 15:12:16 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme1n2 00:06:26.792 No valid GPT data, bailing 00:06:26.792 15:12:16 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme1n2 00:06:26.792 15:12:16 -- scripts/common.sh@394 -- # pt= 00:06:26.792 15:12:16 -- scripts/common.sh@395 -- # return 1 00:06:26.792 15:12:16 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme1n2 bs=1M count=1 00:06:26.792 1+0 records in 00:06:26.792 1+0 records out 00:06:26.792 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00408502 s, 257 MB/s 00:06:26.792 15:12:16 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:06:26.792 15:12:16 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:06:26.792 15:12:16 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme1n3 00:06:26.792 15:12:16 -- scripts/common.sh@381 -- # local block=/dev/nvme1n3 pt 00:06:26.792 15:12:16 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme1n3 00:06:26.792 No valid GPT data, bailing 00:06:26.792 15:12:16 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme1n3 00:06:26.792 15:12:16 -- scripts/common.sh@394 -- # pt= 00:06:26.792 15:12:16 -- scripts/common.sh@395 -- # return 1 00:06:26.792 15:12:16 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme1n3 bs=1M count=1 00:06:26.792 1+0 records in 00:06:26.792 1+0 records out 00:06:26.792 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00650896 s, 161 MB/s 00:06:26.792 15:12:16 -- spdk/autotest.sh@105 -- # sync 00:06:26.792 15:12:16 -- spdk/autotest.sh@107 -- # xtrace_disable_per_cmd reap_spdk_processes 00:06:26.792 15:12:16 -- common/autotest_common.sh@22 -- # eval 'reap_spdk_processes 12> /dev/null' 00:06:26.792 15:12:16 -- common/autotest_common.sh@22 -- # reap_spdk_processes 00:06:29.332 15:12:19 -- spdk/autotest.sh@111 -- # uname -s 00:06:29.332 15:12:19 -- spdk/autotest.sh@111 -- # [[ Linux == Linux ]] 00:06:29.332 15:12:19 -- spdk/autotest.sh@111 -- # [[ 0 -eq 1 ]] 00:06:29.332 15:12:19 -- spdk/autotest.sh@115 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh status 00:06:29.903 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:06:29.903 Hugepages 00:06:29.903 node hugesize free / total 00:06:29.903 node0 1048576kB 0 / 0 00:06:29.903 node0 2048kB 0 / 0 00:06:29.903 00:06:29.903 Type BDF Vendor Device NUMA Driver Device Block devices 00:06:29.903 virtio 0000:00:03.0 1af4 1001 unknown virtio-pci - vda 00:06:30.163 NVMe 0000:00:10.0 1b36 0010 unknown nvme nvme0 nvme0n1 00:06:30.163 NVMe 0000:00:11.0 1b36 0010 unknown nvme nvme1 nvme1n1 nvme1n2 nvme1n3 00:06:30.163 15:12:20 -- spdk/autotest.sh@117 -- # uname -s 00:06:30.163 15:12:20 -- spdk/autotest.sh@117 -- # [[ Linux == Linux ]] 00:06:30.163 15:12:20 -- spdk/autotest.sh@119 -- # nvme_namespace_revert 00:06:30.163 15:12:20 -- common/autotest_common.sh@1516 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh 00:06:31.104 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:06:31.104 0000:00:10.0 (1b36 0010): nvme -> uio_pci_generic 00:06:31.104 0000:00:11.0 (1b36 0010): nvme -> uio_pci_generic 00:06:31.395 15:12:21 -- common/autotest_common.sh@1517 -- # sleep 1 00:06:32.336 15:12:22 -- common/autotest_common.sh@1518 -- # bdfs=() 00:06:32.336 15:12:22 -- common/autotest_common.sh@1518 -- # local bdfs 00:06:32.336 15:12:22 -- common/autotest_common.sh@1520 -- # bdfs=($(get_nvme_bdfs)) 00:06:32.336 15:12:22 -- common/autotest_common.sh@1520 -- # get_nvme_bdfs 00:06:32.336 15:12:22 -- common/autotest_common.sh@1498 -- # bdfs=() 00:06:32.336 15:12:22 -- common/autotest_common.sh@1498 -- # local bdfs 00:06:32.336 15:12:22 -- common/autotest_common.sh@1499 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:06:32.336 15:12:22 -- common/autotest_common.sh@1499 -- # /home/vagrant/spdk_repo/spdk/scripts/gen_nvme.sh 00:06:32.336 15:12:22 -- common/autotest_common.sh@1499 -- # jq -r '.config[].params.traddr' 00:06:32.336 15:12:22 -- common/autotest_common.sh@1500 -- # (( 2 == 0 )) 00:06:32.336 15:12:22 -- common/autotest_common.sh@1504 -- # printf '%s\n' 0000:00:10.0 0000:00:11.0 00:06:32.336 15:12:22 -- common/autotest_common.sh@1522 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh reset 00:06:32.906 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:06:32.906 Waiting for block devices as requested 00:06:32.906 0000:00:11.0 (1b36 0010): uio_pci_generic -> nvme 00:06:32.906 0000:00:10.0 (1b36 0010): uio_pci_generic -> nvme 00:06:33.173 15:12:23 -- common/autotest_common.sh@1524 -- # for bdf in "${bdfs[@]}" 00:06:33.173 15:12:23 -- common/autotest_common.sh@1525 -- # get_nvme_ctrlr_from_bdf 0000:00:10.0 00:06:33.173 15:12:23 -- common/autotest_common.sh@1487 -- # readlink -f /sys/class/nvme/nvme0 /sys/class/nvme/nvme1 00:06:33.173 15:12:23 -- common/autotest_common.sh@1487 -- # grep 0000:00:10.0/nvme/nvme 00:06:33.173 15:12:23 -- common/autotest_common.sh@1487 -- # bdf_sysfs_path=/sys/devices/pci0000:00/0000:00:10.0/nvme/nvme1 00:06:33.173 15:12:23 -- common/autotest_common.sh@1488 -- # [[ -z /sys/devices/pci0000:00/0000:00:10.0/nvme/nvme1 ]] 00:06:33.173 15:12:23 -- common/autotest_common.sh@1492 -- # basename /sys/devices/pci0000:00/0000:00:10.0/nvme/nvme1 00:06:33.173 15:12:23 -- common/autotest_common.sh@1492 -- # printf '%s\n' nvme1 00:06:33.173 15:12:23 -- common/autotest_common.sh@1525 -- # nvme_ctrlr=/dev/nvme1 00:06:33.173 15:12:23 -- common/autotest_common.sh@1526 -- # [[ -z /dev/nvme1 ]] 00:06:33.173 15:12:23 -- common/autotest_common.sh@1531 -- # nvme id-ctrl /dev/nvme1 00:06:33.173 15:12:23 -- common/autotest_common.sh@1531 -- # grep oacs 00:06:33.173 15:12:23 -- common/autotest_common.sh@1531 -- # cut -d: -f2 00:06:33.173 15:12:23 -- common/autotest_common.sh@1531 -- # oacs=' 0x12a' 00:06:33.173 15:12:23 -- common/autotest_common.sh@1532 -- # oacs_ns_manage=8 00:06:33.173 15:12:23 -- common/autotest_common.sh@1534 -- # [[ 8 -ne 0 ]] 00:06:33.173 15:12:23 -- common/autotest_common.sh@1540 -- # nvme id-ctrl /dev/nvme1 00:06:33.173 15:12:23 -- common/autotest_common.sh@1540 -- # grep unvmcap 00:06:33.173 15:12:23 -- common/autotest_common.sh@1540 -- # cut -d: -f2 00:06:33.173 15:12:23 -- common/autotest_common.sh@1540 -- # unvmcap=' 0' 00:06:33.173 15:12:23 -- common/autotest_common.sh@1541 -- # [[ 0 -eq 0 ]] 00:06:33.173 15:12:23 -- common/autotest_common.sh@1543 -- # continue 00:06:33.173 15:12:23 -- common/autotest_common.sh@1524 -- # for bdf in "${bdfs[@]}" 00:06:33.173 15:12:23 -- common/autotest_common.sh@1525 -- # get_nvme_ctrlr_from_bdf 0000:00:11.0 00:06:33.173 15:12:23 -- common/autotest_common.sh@1487 -- # readlink -f /sys/class/nvme/nvme0 /sys/class/nvme/nvme1 00:06:33.173 15:12:23 -- common/autotest_common.sh@1487 -- # grep 0000:00:11.0/nvme/nvme 00:06:33.173 15:12:23 -- common/autotest_common.sh@1487 -- # bdf_sysfs_path=/sys/devices/pci0000:00/0000:00:11.0/nvme/nvme0 00:06:33.173 15:12:23 -- common/autotest_common.sh@1488 -- # [[ -z /sys/devices/pci0000:00/0000:00:11.0/nvme/nvme0 ]] 00:06:33.174 15:12:23 -- common/autotest_common.sh@1492 -- # basename /sys/devices/pci0000:00/0000:00:11.0/nvme/nvme0 00:06:33.174 15:12:23 -- common/autotest_common.sh@1492 -- # printf '%s\n' nvme0 00:06:33.174 15:12:23 -- common/autotest_common.sh@1525 -- # nvme_ctrlr=/dev/nvme0 00:06:33.174 15:12:23 -- common/autotest_common.sh@1526 -- # [[ -z /dev/nvme0 ]] 00:06:33.174 15:12:23 -- common/autotest_common.sh@1531 -- # nvme id-ctrl /dev/nvme0 00:06:33.174 15:12:23 -- common/autotest_common.sh@1531 -- # grep oacs 00:06:33.174 15:12:23 -- common/autotest_common.sh@1531 -- # cut -d: -f2 00:06:33.174 15:12:23 -- common/autotest_common.sh@1531 -- # oacs=' 0x12a' 00:06:33.174 15:12:23 -- common/autotest_common.sh@1532 -- # oacs_ns_manage=8 00:06:33.174 15:12:23 -- common/autotest_common.sh@1534 -- # [[ 8 -ne 0 ]] 00:06:33.174 15:12:23 -- common/autotest_common.sh@1540 -- # nvme id-ctrl /dev/nvme0 00:06:33.174 15:12:23 -- common/autotest_common.sh@1540 -- # grep unvmcap 00:06:33.174 15:12:23 -- common/autotest_common.sh@1540 -- # cut -d: -f2 00:06:33.174 15:12:23 -- common/autotest_common.sh@1540 -- # unvmcap=' 0' 00:06:33.174 15:12:23 -- common/autotest_common.sh@1541 -- # [[ 0 -eq 0 ]] 00:06:33.174 15:12:23 -- common/autotest_common.sh@1543 -- # continue 00:06:33.174 15:12:23 -- spdk/autotest.sh@122 -- # timing_exit pre_cleanup 00:06:33.174 15:12:23 -- common/autotest_common.sh@732 -- # xtrace_disable 00:06:33.174 15:12:23 -- common/autotest_common.sh@10 -- # set +x 00:06:33.174 15:12:23 -- spdk/autotest.sh@125 -- # timing_enter afterboot 00:06:33.174 15:12:23 -- common/autotest_common.sh@726 -- # xtrace_disable 00:06:33.174 15:12:23 -- common/autotest_common.sh@10 -- # set +x 00:06:33.174 15:12:23 -- spdk/autotest.sh@126 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh 00:06:34.139 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:06:34.139 0000:00:10.0 (1b36 0010): nvme -> uio_pci_generic 00:06:34.139 0000:00:11.0 (1b36 0010): nvme -> uio_pci_generic 00:06:34.399 15:12:24 -- spdk/autotest.sh@127 -- # timing_exit afterboot 00:06:34.399 15:12:24 -- common/autotest_common.sh@732 -- # xtrace_disable 00:06:34.399 15:12:24 -- common/autotest_common.sh@10 -- # set +x 00:06:34.399 15:12:24 -- spdk/autotest.sh@131 -- # opal_revert_cleanup 00:06:34.399 15:12:24 -- common/autotest_common.sh@1578 -- # mapfile -t bdfs 00:06:34.399 15:12:24 -- common/autotest_common.sh@1578 -- # get_nvme_bdfs_by_id 0x0a54 00:06:34.399 15:12:24 -- common/autotest_common.sh@1563 -- # bdfs=() 00:06:34.399 15:12:24 -- common/autotest_common.sh@1563 -- # _bdfs=() 00:06:34.399 15:12:24 -- common/autotest_common.sh@1563 -- # local bdfs _bdfs 00:06:34.399 15:12:24 -- common/autotest_common.sh@1564 -- # _bdfs=($(get_nvme_bdfs)) 00:06:34.399 15:12:24 -- common/autotest_common.sh@1564 -- # get_nvme_bdfs 00:06:34.399 15:12:24 -- common/autotest_common.sh@1498 -- # bdfs=() 00:06:34.399 15:12:24 -- common/autotest_common.sh@1498 -- # local bdfs 00:06:34.399 15:12:24 -- common/autotest_common.sh@1499 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:06:34.399 15:12:24 -- common/autotest_common.sh@1499 -- # /home/vagrant/spdk_repo/spdk/scripts/gen_nvme.sh 00:06:34.399 15:12:24 -- common/autotest_common.sh@1499 -- # jq -r '.config[].params.traddr' 00:06:34.399 15:12:24 -- common/autotest_common.sh@1500 -- # (( 2 == 0 )) 00:06:34.399 15:12:24 -- common/autotest_common.sh@1504 -- # printf '%s\n' 0000:00:10.0 0000:00:11.0 00:06:34.399 15:12:24 -- common/autotest_common.sh@1565 -- # for bdf in "${_bdfs[@]}" 00:06:34.399 15:12:24 -- common/autotest_common.sh@1566 -- # cat /sys/bus/pci/devices/0000:00:10.0/device 00:06:34.399 15:12:24 -- common/autotest_common.sh@1566 -- # device=0x0010 00:06:34.399 15:12:24 -- common/autotest_common.sh@1567 -- # [[ 0x0010 == \0\x\0\a\5\4 ]] 00:06:34.399 15:12:24 -- common/autotest_common.sh@1565 -- # for bdf in "${_bdfs[@]}" 00:06:34.399 15:12:24 -- common/autotest_common.sh@1566 -- # cat /sys/bus/pci/devices/0000:00:11.0/device 00:06:34.399 15:12:24 -- common/autotest_common.sh@1566 -- # device=0x0010 00:06:34.399 15:12:24 -- common/autotest_common.sh@1567 -- # [[ 0x0010 == \0\x\0\a\5\4 ]] 00:06:34.399 15:12:24 -- common/autotest_common.sh@1572 -- # (( 0 > 0 )) 00:06:34.399 15:12:24 -- common/autotest_common.sh@1572 -- # return 0 00:06:34.399 15:12:24 -- common/autotest_common.sh@1579 -- # [[ -z '' ]] 00:06:34.399 15:12:24 -- common/autotest_common.sh@1580 -- # return 0 00:06:34.399 15:12:24 -- spdk/autotest.sh@137 -- # '[' 0 -eq 1 ']' 00:06:34.399 15:12:24 -- spdk/autotest.sh@141 -- # '[' 1 -eq 1 ']' 00:06:34.399 15:12:24 -- spdk/autotest.sh@142 -- # [[ 0 -eq 1 ]] 00:06:34.399 15:12:24 -- spdk/autotest.sh@142 -- # [[ 0 -eq 1 ]] 00:06:34.399 15:12:24 -- spdk/autotest.sh@149 -- # timing_enter lib 00:06:34.399 15:12:24 -- common/autotest_common.sh@726 -- # xtrace_disable 00:06:34.399 15:12:24 -- common/autotest_common.sh@10 -- # set +x 00:06:34.399 15:12:24 -- spdk/autotest.sh@151 -- # [[ 0 -eq 1 ]] 00:06:34.399 15:12:24 -- spdk/autotest.sh@155 -- # run_test env /home/vagrant/spdk_repo/spdk/test/env/env.sh 00:06:34.399 15:12:24 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:34.399 15:12:24 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:34.399 15:12:24 -- common/autotest_common.sh@10 -- # set +x 00:06:34.399 ************************************ 00:06:34.399 START TEST env 00:06:34.399 ************************************ 00:06:34.399 15:12:24 env -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/env/env.sh 00:06:34.659 * Looking for test storage... 00:06:34.659 * Found test storage at /home/vagrant/spdk_repo/spdk/test/env 00:06:34.659 15:12:24 env -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:06:34.659 15:12:24 env -- common/autotest_common.sh@1693 -- # lcov --version 00:06:34.659 15:12:24 env -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:06:34.659 15:12:24 env -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:06:34.659 15:12:24 env -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:34.659 15:12:24 env -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:34.659 15:12:24 env -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:34.659 15:12:24 env -- scripts/common.sh@336 -- # IFS=.-: 00:06:34.659 15:12:24 env -- scripts/common.sh@336 -- # read -ra ver1 00:06:34.659 15:12:24 env -- scripts/common.sh@337 -- # IFS=.-: 00:06:34.659 15:12:24 env -- scripts/common.sh@337 -- # read -ra ver2 00:06:34.659 15:12:24 env -- scripts/common.sh@338 -- # local 'op=<' 00:06:34.659 15:12:24 env -- scripts/common.sh@340 -- # ver1_l=2 00:06:34.659 15:12:24 env -- scripts/common.sh@341 -- # ver2_l=1 00:06:34.659 15:12:24 env -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:34.659 15:12:24 env -- scripts/common.sh@344 -- # case "$op" in 00:06:34.659 15:12:24 env -- scripts/common.sh@345 -- # : 1 00:06:34.659 15:12:24 env -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:34.659 15:12:24 env -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:34.659 15:12:24 env -- scripts/common.sh@365 -- # decimal 1 00:06:34.659 15:12:24 env -- scripts/common.sh@353 -- # local d=1 00:06:34.659 15:12:24 env -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:34.659 15:12:24 env -- scripts/common.sh@355 -- # echo 1 00:06:34.659 15:12:24 env -- scripts/common.sh@365 -- # ver1[v]=1 00:06:34.660 15:12:24 env -- scripts/common.sh@366 -- # decimal 2 00:06:34.660 15:12:24 env -- scripts/common.sh@353 -- # local d=2 00:06:34.660 15:12:24 env -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:34.660 15:12:24 env -- scripts/common.sh@355 -- # echo 2 00:06:34.660 15:12:24 env -- scripts/common.sh@366 -- # ver2[v]=2 00:06:34.660 15:12:24 env -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:34.660 15:12:24 env -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:34.660 15:12:24 env -- scripts/common.sh@368 -- # return 0 00:06:34.660 15:12:24 env -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:34.660 15:12:24 env -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:06:34.660 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:34.660 --rc genhtml_branch_coverage=1 00:06:34.660 --rc genhtml_function_coverage=1 00:06:34.660 --rc genhtml_legend=1 00:06:34.660 --rc geninfo_all_blocks=1 00:06:34.660 --rc geninfo_unexecuted_blocks=1 00:06:34.660 00:06:34.660 ' 00:06:34.660 15:12:24 env -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:06:34.660 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:34.660 --rc genhtml_branch_coverage=1 00:06:34.660 --rc genhtml_function_coverage=1 00:06:34.660 --rc genhtml_legend=1 00:06:34.660 --rc geninfo_all_blocks=1 00:06:34.660 --rc geninfo_unexecuted_blocks=1 00:06:34.660 00:06:34.660 ' 00:06:34.660 15:12:24 env -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:06:34.660 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:34.660 --rc genhtml_branch_coverage=1 00:06:34.660 --rc genhtml_function_coverage=1 00:06:34.660 --rc genhtml_legend=1 00:06:34.660 --rc geninfo_all_blocks=1 00:06:34.660 --rc geninfo_unexecuted_blocks=1 00:06:34.660 00:06:34.660 ' 00:06:34.660 15:12:24 env -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:06:34.660 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:34.660 --rc genhtml_branch_coverage=1 00:06:34.660 --rc genhtml_function_coverage=1 00:06:34.660 --rc genhtml_legend=1 00:06:34.660 --rc geninfo_all_blocks=1 00:06:34.660 --rc geninfo_unexecuted_blocks=1 00:06:34.660 00:06:34.660 ' 00:06:34.660 15:12:24 env -- env/env.sh@10 -- # run_test env_memory /home/vagrant/spdk_repo/spdk/test/env/memory/memory_ut 00:06:34.660 15:12:24 env -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:34.660 15:12:24 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:34.660 15:12:24 env -- common/autotest_common.sh@10 -- # set +x 00:06:34.660 ************************************ 00:06:34.660 START TEST env_memory 00:06:34.660 ************************************ 00:06:34.660 15:12:24 env.env_memory -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/env/memory/memory_ut 00:06:34.660 00:06:34.660 00:06:34.660 CUnit - A unit testing framework for C - Version 2.1-3 00:06:34.660 http://cunit.sourceforge.net/ 00:06:34.660 00:06:34.660 00:06:34.660 Suite: memory 00:06:34.920 Test: alloc and free memory map ...[2024-11-19 15:12:25.000421] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 283:spdk_mem_map_alloc: *ERROR*: Initial mem_map notify failed 00:06:34.920 passed 00:06:34.920 Test: mem map translation ...[2024-11-19 15:12:25.041789] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 595:spdk_mem_map_set_translation: *ERROR*: invalid spdk_mem_map_set_translation parameters, vaddr=2097152 len=1234 00:06:34.920 [2024-11-19 15:12:25.041930] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 595:spdk_mem_map_set_translation: *ERROR*: invalid spdk_mem_map_set_translation parameters, vaddr=1234 len=2097152 00:06:34.920 [2024-11-19 15:12:25.042089] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 589:spdk_mem_map_set_translation: *ERROR*: invalid usermode virtual address 281474976710656 00:06:34.920 [2024-11-19 15:12:25.042203] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 605:spdk_mem_map_set_translation: *ERROR*: could not get 0xffffffe00000 map 00:06:34.920 passed 00:06:34.920 Test: mem map registration ...[2024-11-19 15:12:25.106345] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 347:spdk_mem_register: *ERROR*: invalid spdk_mem_register parameters, vaddr=200000 len=1234 00:06:34.920 [2024-11-19 15:12:25.106469] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 347:spdk_mem_register: *ERROR*: invalid spdk_mem_register parameters, vaddr=4d2 len=2097152 00:06:34.920 passed 00:06:34.920 Test: mem map adjacent registrations ...passed 00:06:34.920 00:06:34.920 Run Summary: Type Total Ran Passed Failed Inactive 00:06:34.920 suites 1 1 n/a 0 0 00:06:34.920 tests 4 4 4 0 0 00:06:34.920 asserts 152 152 152 0 n/a 00:06:34.920 00:06:34.920 Elapsed time = 0.228 seconds 00:06:34.920 00:06:34.920 real 0m0.276s 00:06:34.920 user 0m0.244s 00:06:34.920 sys 0m0.022s 00:06:34.920 15:12:25 env.env_memory -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:34.920 ************************************ 00:06:34.920 END TEST env_memory 00:06:34.920 ************************************ 00:06:34.920 15:12:25 env.env_memory -- common/autotest_common.sh@10 -- # set +x 00:06:35.180 15:12:25 env -- env/env.sh@11 -- # run_test env_vtophys /home/vagrant/spdk_repo/spdk/test/env/vtophys/vtophys 00:06:35.180 15:12:25 env -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:35.180 15:12:25 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:35.180 15:12:25 env -- common/autotest_common.sh@10 -- # set +x 00:06:35.180 ************************************ 00:06:35.180 START TEST env_vtophys 00:06:35.180 ************************************ 00:06:35.180 15:12:25 env.env_vtophys -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/env/vtophys/vtophys 00:06:35.180 EAL: lib.eal log level changed from notice to debug 00:06:35.180 EAL: Detected lcore 0 as core 0 on socket 0 00:06:35.180 EAL: Detected lcore 1 as core 0 on socket 0 00:06:35.180 EAL: Detected lcore 2 as core 0 on socket 0 00:06:35.180 EAL: Detected lcore 3 as core 0 on socket 0 00:06:35.180 EAL: Detected lcore 4 as core 0 on socket 0 00:06:35.180 EAL: Detected lcore 5 as core 0 on socket 0 00:06:35.180 EAL: Detected lcore 6 as core 0 on socket 0 00:06:35.180 EAL: Detected lcore 7 as core 0 on socket 0 00:06:35.180 EAL: Detected lcore 8 as core 0 on socket 0 00:06:35.180 EAL: Detected lcore 9 as core 0 on socket 0 00:06:35.180 EAL: Maximum logical cores by configuration: 128 00:06:35.180 EAL: Detected CPU lcores: 10 00:06:35.180 EAL: Detected NUMA nodes: 1 00:06:35.180 EAL: Checking presence of .so 'librte_eal.so.23.0' 00:06:35.180 EAL: Detected shared linkage of DPDK 00:06:35.180 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_pci.so.23.0 00:06:35.180 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_vdev.so.23.0 00:06:35.180 EAL: Registered [vdev] bus. 00:06:35.180 EAL: bus.vdev log level changed from disabled to notice 00:06:35.180 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_mempool_ring.so.23.0 00:06:35.180 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_net_i40e.so.23.0 00:06:35.180 EAL: pmd.net.i40e.init log level changed from disabled to notice 00:06:35.180 EAL: pmd.net.i40e.driver log level changed from disabled to notice 00:06:35.180 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_pci.so 00:06:35.180 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_vdev.so 00:06:35.180 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_mempool_ring.so 00:06:35.180 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_net_i40e.so 00:06:35.180 EAL: No shared files mode enabled, IPC will be disabled 00:06:35.180 EAL: No shared files mode enabled, IPC is disabled 00:06:35.180 EAL: Selected IOVA mode 'PA' 00:06:35.180 EAL: Probing VFIO support... 00:06:35.180 EAL: Module /sys/module/vfio not found! error 2 (No such file or directory) 00:06:35.180 EAL: VFIO modules not loaded, skipping VFIO support... 00:06:35.180 EAL: Ask a virtual area of 0x2e000 bytes 00:06:35.180 EAL: Virtual area found at 0x200000000000 (size = 0x2e000) 00:06:35.180 EAL: Setting up physically contiguous memory... 00:06:35.180 EAL: Setting maximum number of open files to 524288 00:06:35.180 EAL: Detected memory type: socket_id:0 hugepage_sz:2097152 00:06:35.180 EAL: Creating 4 segment lists: n_segs:8192 socket_id:0 hugepage_sz:2097152 00:06:35.180 EAL: Ask a virtual area of 0x61000 bytes 00:06:35.180 EAL: Virtual area found at 0x20000002e000 (size = 0x61000) 00:06:35.180 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:06:35.180 EAL: Ask a virtual area of 0x400000000 bytes 00:06:35.180 EAL: Virtual area found at 0x200000200000 (size = 0x400000000) 00:06:35.180 EAL: VA reserved for memseg list at 0x200000200000, size 400000000 00:06:35.180 EAL: Ask a virtual area of 0x61000 bytes 00:06:35.180 EAL: Virtual area found at 0x200400200000 (size = 0x61000) 00:06:35.180 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:06:35.180 EAL: Ask a virtual area of 0x400000000 bytes 00:06:35.180 EAL: Virtual area found at 0x200400400000 (size = 0x400000000) 00:06:35.180 EAL: VA reserved for memseg list at 0x200400400000, size 400000000 00:06:35.180 EAL: Ask a virtual area of 0x61000 bytes 00:06:35.180 EAL: Virtual area found at 0x200800400000 (size = 0x61000) 00:06:35.180 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:06:35.180 EAL: Ask a virtual area of 0x400000000 bytes 00:06:35.180 EAL: Virtual area found at 0x200800600000 (size = 0x400000000) 00:06:35.180 EAL: VA reserved for memseg list at 0x200800600000, size 400000000 00:06:35.180 EAL: Ask a virtual area of 0x61000 bytes 00:06:35.180 EAL: Virtual area found at 0x200c00600000 (size = 0x61000) 00:06:35.180 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:06:35.180 EAL: Ask a virtual area of 0x400000000 bytes 00:06:35.180 EAL: Virtual area found at 0x200c00800000 (size = 0x400000000) 00:06:35.180 EAL: VA reserved for memseg list at 0x200c00800000, size 400000000 00:06:35.180 EAL: Hugepages will be freed exactly as allocated. 00:06:35.180 EAL: No shared files mode enabled, IPC is disabled 00:06:35.180 EAL: No shared files mode enabled, IPC is disabled 00:06:35.180 EAL: TSC frequency is ~2290000 KHz 00:06:35.180 EAL: Main lcore 0 is ready (tid=7fd9c96fba40;cpuset=[0]) 00:06:35.180 EAL: Trying to obtain current memory policy. 00:06:35.180 EAL: Setting policy MPOL_PREFERRED for socket 0 00:06:35.180 EAL: Restoring previous memory policy: 0 00:06:35.180 EAL: request: mp_malloc_sync 00:06:35.180 EAL: No shared files mode enabled, IPC is disabled 00:06:35.180 EAL: Heap on socket 0 was expanded by 2MB 00:06:35.180 EAL: Module /sys/module/vfio not found! error 2 (No such file or directory) 00:06:35.180 EAL: No shared files mode enabled, IPC is disabled 00:06:35.180 EAL: No PCI address specified using 'addr=' in: bus=pci 00:06:35.180 EAL: Mem event callback 'spdk:(nil)' registered 00:06:35.180 EAL: Module /sys/module/vfio_pci not found! error 2 (No such file or directory) 00:06:35.180 00:06:35.180 00:06:35.180 CUnit - A unit testing framework for C - Version 2.1-3 00:06:35.180 http://cunit.sourceforge.net/ 00:06:35.180 00:06:35.180 00:06:35.180 Suite: components_suite 00:06:35.751 Test: vtophys_malloc_test ...passed 00:06:35.751 Test: vtophys_spdk_malloc_test ...EAL: Trying to obtain current memory policy. 00:06:35.751 EAL: Setting policy MPOL_PREFERRED for socket 0 00:06:35.751 EAL: Restoring previous memory policy: 4 00:06:35.751 EAL: Calling mem event callback 'spdk:(nil)' 00:06:35.751 EAL: request: mp_malloc_sync 00:06:35.751 EAL: No shared files mode enabled, IPC is disabled 00:06:35.751 EAL: Heap on socket 0 was expanded by 4MB 00:06:35.751 EAL: Calling mem event callback 'spdk:(nil)' 00:06:35.751 EAL: request: mp_malloc_sync 00:06:35.751 EAL: No shared files mode enabled, IPC is disabled 00:06:35.751 EAL: Heap on socket 0 was shrunk by 4MB 00:06:35.751 EAL: Trying to obtain current memory policy. 00:06:35.751 EAL: Setting policy MPOL_PREFERRED for socket 0 00:06:35.751 EAL: Restoring previous memory policy: 4 00:06:35.751 EAL: Calling mem event callback 'spdk:(nil)' 00:06:35.751 EAL: request: mp_malloc_sync 00:06:35.751 EAL: No shared files mode enabled, IPC is disabled 00:06:35.751 EAL: Heap on socket 0 was expanded by 6MB 00:06:35.751 EAL: Calling mem event callback 'spdk:(nil)' 00:06:35.751 EAL: request: mp_malloc_sync 00:06:35.751 EAL: No shared files mode enabled, IPC is disabled 00:06:35.751 EAL: Heap on socket 0 was shrunk by 6MB 00:06:35.751 EAL: Trying to obtain current memory policy. 00:06:35.751 EAL: Setting policy MPOL_PREFERRED for socket 0 00:06:35.751 EAL: Restoring previous memory policy: 4 00:06:35.751 EAL: Calling mem event callback 'spdk:(nil)' 00:06:35.751 EAL: request: mp_malloc_sync 00:06:35.751 EAL: No shared files mode enabled, IPC is disabled 00:06:35.751 EAL: Heap on socket 0 was expanded by 10MB 00:06:35.751 EAL: Calling mem event callback 'spdk:(nil)' 00:06:35.751 EAL: request: mp_malloc_sync 00:06:35.751 EAL: No shared files mode enabled, IPC is disabled 00:06:35.751 EAL: Heap on socket 0 was shrunk by 10MB 00:06:35.751 EAL: Trying to obtain current memory policy. 00:06:35.751 EAL: Setting policy MPOL_PREFERRED for socket 0 00:06:35.751 EAL: Restoring previous memory policy: 4 00:06:35.751 EAL: Calling mem event callback 'spdk:(nil)' 00:06:35.751 EAL: request: mp_malloc_sync 00:06:35.751 EAL: No shared files mode enabled, IPC is disabled 00:06:35.751 EAL: Heap on socket 0 was expanded by 18MB 00:06:35.751 EAL: Calling mem event callback 'spdk:(nil)' 00:06:35.751 EAL: request: mp_malloc_sync 00:06:35.751 EAL: No shared files mode enabled, IPC is disabled 00:06:35.751 EAL: Heap on socket 0 was shrunk by 18MB 00:06:35.751 EAL: Trying to obtain current memory policy. 00:06:35.751 EAL: Setting policy MPOL_PREFERRED for socket 0 00:06:35.751 EAL: Restoring previous memory policy: 4 00:06:35.751 EAL: Calling mem event callback 'spdk:(nil)' 00:06:35.751 EAL: request: mp_malloc_sync 00:06:35.751 EAL: No shared files mode enabled, IPC is disabled 00:06:35.751 EAL: Heap on socket 0 was expanded by 34MB 00:06:35.751 EAL: Calling mem event callback 'spdk:(nil)' 00:06:35.751 EAL: request: mp_malloc_sync 00:06:35.751 EAL: No shared files mode enabled, IPC is disabled 00:06:35.751 EAL: Heap on socket 0 was shrunk by 34MB 00:06:35.751 EAL: Trying to obtain current memory policy. 00:06:35.751 EAL: Setting policy MPOL_PREFERRED for socket 0 00:06:35.751 EAL: Restoring previous memory policy: 4 00:06:35.751 EAL: Calling mem event callback 'spdk:(nil)' 00:06:35.751 EAL: request: mp_malloc_sync 00:06:35.751 EAL: No shared files mode enabled, IPC is disabled 00:06:35.751 EAL: Heap on socket 0 was expanded by 66MB 00:06:35.751 EAL: Calling mem event callback 'spdk:(nil)' 00:06:35.751 EAL: request: mp_malloc_sync 00:06:35.751 EAL: No shared files mode enabled, IPC is disabled 00:06:35.751 EAL: Heap on socket 0 was shrunk by 66MB 00:06:35.751 EAL: Trying to obtain current memory policy. 00:06:35.751 EAL: Setting policy MPOL_PREFERRED for socket 0 00:06:35.751 EAL: Restoring previous memory policy: 4 00:06:35.751 EAL: Calling mem event callback 'spdk:(nil)' 00:06:35.751 EAL: request: mp_malloc_sync 00:06:35.751 EAL: No shared files mode enabled, IPC is disabled 00:06:35.751 EAL: Heap on socket 0 was expanded by 130MB 00:06:36.011 EAL: Calling mem event callback 'spdk:(nil)' 00:06:36.011 EAL: request: mp_malloc_sync 00:06:36.011 EAL: No shared files mode enabled, IPC is disabled 00:06:36.011 EAL: Heap on socket 0 was shrunk by 130MB 00:06:36.011 EAL: Trying to obtain current memory policy. 00:06:36.011 EAL: Setting policy MPOL_PREFERRED for socket 0 00:06:36.011 EAL: Restoring previous memory policy: 4 00:06:36.011 EAL: Calling mem event callback 'spdk:(nil)' 00:06:36.011 EAL: request: mp_malloc_sync 00:06:36.011 EAL: No shared files mode enabled, IPC is disabled 00:06:36.011 EAL: Heap on socket 0 was expanded by 258MB 00:06:36.011 EAL: Calling mem event callback 'spdk:(nil)' 00:06:36.271 EAL: request: mp_malloc_sync 00:06:36.271 EAL: No shared files mode enabled, IPC is disabled 00:06:36.271 EAL: Heap on socket 0 was shrunk by 258MB 00:06:36.271 EAL: Trying to obtain current memory policy. 00:06:36.271 EAL: Setting policy MPOL_PREFERRED for socket 0 00:06:36.271 EAL: Restoring previous memory policy: 4 00:06:36.271 EAL: Calling mem event callback 'spdk:(nil)' 00:06:36.271 EAL: request: mp_malloc_sync 00:06:36.271 EAL: No shared files mode enabled, IPC is disabled 00:06:36.271 EAL: Heap on socket 0 was expanded by 514MB 00:06:36.531 EAL: Calling mem event callback 'spdk:(nil)' 00:06:36.790 EAL: request: mp_malloc_sync 00:06:36.790 EAL: No shared files mode enabled, IPC is disabled 00:06:36.791 EAL: Heap on socket 0 was shrunk by 514MB 00:06:36.791 EAL: Trying to obtain current memory policy. 00:06:36.791 EAL: Setting policy MPOL_PREFERRED for socket 0 00:06:37.049 EAL: Restoring previous memory policy: 4 00:06:37.049 EAL: Calling mem event callback 'spdk:(nil)' 00:06:37.049 EAL: request: mp_malloc_sync 00:06:37.049 EAL: No shared files mode enabled, IPC is disabled 00:06:37.049 EAL: Heap on socket 0 was expanded by 1026MB 00:06:37.308 EAL: Calling mem event callback 'spdk:(nil)' 00:06:37.567 passed 00:06:37.567 00:06:37.567 Run Summary: Type Total Ran Passed Failed Inactive 00:06:37.567 suites 1 1 n/a 0 0 00:06:37.567 tests 2 2 2 0 0 00:06:37.567 asserts 5442 5442 5442 0 n/a 00:06:37.567 00:06:37.567 Elapsed time = 2.356 seconds 00:06:37.567 EAL: request: mp_malloc_sync 00:06:37.567 EAL: No shared files mode enabled, IPC is disabled 00:06:37.567 EAL: Heap on socket 0 was shrunk by 1026MB 00:06:37.567 EAL: Calling mem event callback 'spdk:(nil)' 00:06:37.567 EAL: request: mp_malloc_sync 00:06:37.567 EAL: No shared files mode enabled, IPC is disabled 00:06:37.567 EAL: Heap on socket 0 was shrunk by 2MB 00:06:37.567 EAL: No shared files mode enabled, IPC is disabled 00:06:37.567 EAL: No shared files mode enabled, IPC is disabled 00:06:37.567 EAL: No shared files mode enabled, IPC is disabled 00:06:37.567 00:06:37.567 real 0m2.612s 00:06:37.567 user 0m1.382s 00:06:37.567 sys 0m1.087s 00:06:37.567 ************************************ 00:06:37.567 END TEST env_vtophys 00:06:37.567 ************************************ 00:06:37.567 15:12:27 env.env_vtophys -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:37.567 15:12:27 env.env_vtophys -- common/autotest_common.sh@10 -- # set +x 00:06:37.827 15:12:27 env -- env/env.sh@12 -- # run_test env_pci /home/vagrant/spdk_repo/spdk/test/env/pci/pci_ut 00:06:37.827 15:12:27 env -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:37.827 15:12:27 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:37.827 15:12:27 env -- common/autotest_common.sh@10 -- # set +x 00:06:37.827 ************************************ 00:06:37.827 START TEST env_pci 00:06:37.827 ************************************ 00:06:37.827 15:12:27 env.env_pci -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/env/pci/pci_ut 00:06:37.827 00:06:37.827 00:06:37.827 CUnit - A unit testing framework for C - Version 2.1-3 00:06:37.827 http://cunit.sourceforge.net/ 00:06:37.827 00:06:37.827 00:06:37.827 Suite: pci 00:06:37.827 Test: pci_hook ...[2024-11-19 15:12:27.993433] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/pci.c:1117:spdk_pci_device_claim: *ERROR*: Cannot create lock on device /var/tmp/spdk_pci_lock_10000:00:01.0, probably process 68881 has claimed it 00:06:37.827 EAL: Cannot find device (10000:00:01.0) 00:06:37.827 EAL: Failed to attach device on primary process 00:06:37.827 passed 00:06:37.827 00:06:37.827 Run Summary: Type Total Ran Passed Failed Inactive 00:06:37.827 suites 1 1 n/a 0 0 00:06:37.827 tests 1 1 1 0 0 00:06:37.827 asserts 25 25 25 0 n/a 00:06:37.827 00:06:37.827 Elapsed time = 0.005 seconds 00:06:37.827 00:06:37.827 real 0m0.091s 00:06:37.827 user 0m0.042s 00:06:37.827 sys 0m0.047s 00:06:37.827 15:12:28 env.env_pci -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:37.827 ************************************ 00:06:37.827 END TEST env_pci 00:06:37.827 ************************************ 00:06:37.827 15:12:28 env.env_pci -- common/autotest_common.sh@10 -- # set +x 00:06:37.827 15:12:28 env -- env/env.sh@14 -- # argv='-c 0x1 ' 00:06:37.827 15:12:28 env -- env/env.sh@15 -- # uname 00:06:37.827 15:12:28 env -- env/env.sh@15 -- # '[' Linux = Linux ']' 00:06:37.827 15:12:28 env -- env/env.sh@22 -- # argv+=--base-virtaddr=0x200000000000 00:06:37.827 15:12:28 env -- env/env.sh@24 -- # run_test env_dpdk_post_init /home/vagrant/spdk_repo/spdk/test/env/env_dpdk_post_init/env_dpdk_post_init -c 0x1 --base-virtaddr=0x200000000000 00:06:37.827 15:12:28 env -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:06:37.827 15:12:28 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:37.827 15:12:28 env -- common/autotest_common.sh@10 -- # set +x 00:06:37.827 ************************************ 00:06:37.827 START TEST env_dpdk_post_init 00:06:37.827 ************************************ 00:06:37.827 15:12:28 env.env_dpdk_post_init -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/env/env_dpdk_post_init/env_dpdk_post_init -c 0x1 --base-virtaddr=0x200000000000 00:06:38.086 EAL: Detected CPU lcores: 10 00:06:38.086 EAL: Detected NUMA nodes: 1 00:06:38.086 EAL: Detected shared linkage of DPDK 00:06:38.086 EAL: Multi-process socket /var/run/dpdk/rte/mp_socket 00:06:38.086 EAL: Selected IOVA mode 'PA' 00:06:38.086 TELEMETRY: No legacy callbacks, legacy socket not created 00:06:38.086 EAL: Probe PCI driver: spdk_nvme (1b36:0010) device: 0000:00:10.0 (socket -1) 00:06:38.086 EAL: Probe PCI driver: spdk_nvme (1b36:0010) device: 0000:00:11.0 (socket -1) 00:06:38.086 Starting DPDK initialization... 00:06:38.086 Starting SPDK post initialization... 00:06:38.086 SPDK NVMe probe 00:06:38.086 Attaching to 0000:00:10.0 00:06:38.086 Attaching to 0000:00:11.0 00:06:38.086 Attached to 0000:00:10.0 00:06:38.086 Attached to 0000:00:11.0 00:06:38.086 Cleaning up... 00:06:38.086 00:06:38.086 real 0m0.249s 00:06:38.086 user 0m0.078s 00:06:38.086 sys 0m0.071s 00:06:38.087 15:12:28 env.env_dpdk_post_init -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:38.087 15:12:28 env.env_dpdk_post_init -- common/autotest_common.sh@10 -- # set +x 00:06:38.087 ************************************ 00:06:38.087 END TEST env_dpdk_post_init 00:06:38.087 ************************************ 00:06:38.087 15:12:28 env -- env/env.sh@26 -- # uname 00:06:38.346 15:12:28 env -- env/env.sh@26 -- # '[' Linux = Linux ']' 00:06:38.346 15:12:28 env -- env/env.sh@29 -- # run_test env_mem_callbacks /home/vagrant/spdk_repo/spdk/test/env/mem_callbacks/mem_callbacks 00:06:38.347 15:12:28 env -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:38.347 15:12:28 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:38.347 15:12:28 env -- common/autotest_common.sh@10 -- # set +x 00:06:38.347 ************************************ 00:06:38.347 START TEST env_mem_callbacks 00:06:38.347 ************************************ 00:06:38.347 15:12:28 env.env_mem_callbacks -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/env/mem_callbacks/mem_callbacks 00:06:38.347 EAL: Detected CPU lcores: 10 00:06:38.347 EAL: Detected NUMA nodes: 1 00:06:38.347 EAL: Detected shared linkage of DPDK 00:06:38.347 EAL: Multi-process socket /var/run/dpdk/rte/mp_socket 00:06:38.347 EAL: Selected IOVA mode 'PA' 00:06:38.347 TELEMETRY: No legacy callbacks, legacy socket not created 00:06:38.347 00:06:38.347 00:06:38.347 CUnit - A unit testing framework for C - Version 2.1-3 00:06:38.347 http://cunit.sourceforge.net/ 00:06:38.347 00:06:38.347 00:06:38.347 Suite: memory 00:06:38.347 Test: test ... 00:06:38.347 register 0x200000200000 2097152 00:06:38.347 malloc 3145728 00:06:38.347 register 0x200000400000 4194304 00:06:38.347 buf 0x200000500000 len 3145728 PASSED 00:06:38.347 malloc 64 00:06:38.347 buf 0x2000004fff40 len 64 PASSED 00:06:38.347 malloc 4194304 00:06:38.347 register 0x200000800000 6291456 00:06:38.347 buf 0x200000a00000 len 4194304 PASSED 00:06:38.347 free 0x200000500000 3145728 00:06:38.347 free 0x2000004fff40 64 00:06:38.347 unregister 0x200000400000 4194304 PASSED 00:06:38.347 free 0x200000a00000 4194304 00:06:38.347 unregister 0x200000800000 6291456 PASSED 00:06:38.347 malloc 8388608 00:06:38.347 register 0x200000400000 10485760 00:06:38.347 buf 0x200000600000 len 8388608 PASSED 00:06:38.347 free 0x200000600000 8388608 00:06:38.347 unregister 0x200000400000 10485760 PASSED 00:06:38.347 passed 00:06:38.347 00:06:38.347 Run Summary: Type Total Ran Passed Failed Inactive 00:06:38.347 suites 1 1 n/a 0 0 00:06:38.347 tests 1 1 1 0 0 00:06:38.347 asserts 15 15 15 0 n/a 00:06:38.347 00:06:38.347 Elapsed time = 0.012 seconds 00:06:38.347 00:06:38.347 real 0m0.181s 00:06:38.347 user 0m0.025s 00:06:38.347 sys 0m0.056s 00:06:38.347 15:12:28 env.env_mem_callbacks -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:38.347 15:12:28 env.env_mem_callbacks -- common/autotest_common.sh@10 -- # set +x 00:06:38.347 ************************************ 00:06:38.347 END TEST env_mem_callbacks 00:06:38.347 ************************************ 00:06:38.347 00:06:38.347 real 0m3.984s 00:06:38.347 user 0m1.999s 00:06:38.347 sys 0m1.654s 00:06:38.347 15:12:28 env -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:38.347 15:12:28 env -- common/autotest_common.sh@10 -- # set +x 00:06:38.347 ************************************ 00:06:38.347 END TEST env 00:06:38.347 ************************************ 00:06:38.606 15:12:28 -- spdk/autotest.sh@156 -- # run_test rpc /home/vagrant/spdk_repo/spdk/test/rpc/rpc.sh 00:06:38.606 15:12:28 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:38.606 15:12:28 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:38.606 15:12:28 -- common/autotest_common.sh@10 -- # set +x 00:06:38.606 ************************************ 00:06:38.606 START TEST rpc 00:06:38.606 ************************************ 00:06:38.606 15:12:28 rpc -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/rpc/rpc.sh 00:06:38.606 * Looking for test storage... 00:06:38.606 * Found test storage at /home/vagrant/spdk_repo/spdk/test/rpc 00:06:38.606 15:12:28 rpc -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:06:38.606 15:12:28 rpc -- common/autotest_common.sh@1693 -- # lcov --version 00:06:38.606 15:12:28 rpc -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:06:38.606 15:12:28 rpc -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:06:38.606 15:12:28 rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:38.606 15:12:28 rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:38.606 15:12:28 rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:38.606 15:12:28 rpc -- scripts/common.sh@336 -- # IFS=.-: 00:06:38.606 15:12:28 rpc -- scripts/common.sh@336 -- # read -ra ver1 00:06:38.607 15:12:28 rpc -- scripts/common.sh@337 -- # IFS=.-: 00:06:38.607 15:12:28 rpc -- scripts/common.sh@337 -- # read -ra ver2 00:06:38.607 15:12:28 rpc -- scripts/common.sh@338 -- # local 'op=<' 00:06:38.607 15:12:28 rpc -- scripts/common.sh@340 -- # ver1_l=2 00:06:38.607 15:12:28 rpc -- scripts/common.sh@341 -- # ver2_l=1 00:06:38.607 15:12:28 rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:38.607 15:12:28 rpc -- scripts/common.sh@344 -- # case "$op" in 00:06:38.607 15:12:28 rpc -- scripts/common.sh@345 -- # : 1 00:06:38.607 15:12:28 rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:38.607 15:12:28 rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:38.607 15:12:28 rpc -- scripts/common.sh@365 -- # decimal 1 00:06:38.607 15:12:28 rpc -- scripts/common.sh@353 -- # local d=1 00:06:38.607 15:12:28 rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:38.607 15:12:28 rpc -- scripts/common.sh@355 -- # echo 1 00:06:38.867 15:12:28 rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:06:38.867 15:12:28 rpc -- scripts/common.sh@366 -- # decimal 2 00:06:38.867 15:12:28 rpc -- scripts/common.sh@353 -- # local d=2 00:06:38.867 15:12:28 rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:38.867 15:12:28 rpc -- scripts/common.sh@355 -- # echo 2 00:06:38.867 15:12:28 rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:06:38.867 15:12:28 rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:38.867 15:12:28 rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:38.867 15:12:28 rpc -- scripts/common.sh@368 -- # return 0 00:06:38.867 15:12:28 rpc -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:38.867 15:12:28 rpc -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:06:38.867 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:38.867 --rc genhtml_branch_coverage=1 00:06:38.867 --rc genhtml_function_coverage=1 00:06:38.867 --rc genhtml_legend=1 00:06:38.867 --rc geninfo_all_blocks=1 00:06:38.867 --rc geninfo_unexecuted_blocks=1 00:06:38.867 00:06:38.867 ' 00:06:38.867 15:12:28 rpc -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:06:38.867 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:38.867 --rc genhtml_branch_coverage=1 00:06:38.867 --rc genhtml_function_coverage=1 00:06:38.867 --rc genhtml_legend=1 00:06:38.867 --rc geninfo_all_blocks=1 00:06:38.867 --rc geninfo_unexecuted_blocks=1 00:06:38.867 00:06:38.867 ' 00:06:38.867 15:12:28 rpc -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:06:38.867 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:38.867 --rc genhtml_branch_coverage=1 00:06:38.867 --rc genhtml_function_coverage=1 00:06:38.867 --rc genhtml_legend=1 00:06:38.867 --rc geninfo_all_blocks=1 00:06:38.867 --rc geninfo_unexecuted_blocks=1 00:06:38.867 00:06:38.867 ' 00:06:38.867 15:12:28 rpc -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:06:38.867 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:38.867 --rc genhtml_branch_coverage=1 00:06:38.867 --rc genhtml_function_coverage=1 00:06:38.867 --rc genhtml_legend=1 00:06:38.867 --rc geninfo_all_blocks=1 00:06:38.867 --rc geninfo_unexecuted_blocks=1 00:06:38.867 00:06:38.867 ' 00:06:38.867 15:12:28 rpc -- rpc/rpc.sh@65 -- # spdk_pid=69008 00:06:38.867 15:12:28 rpc -- rpc/rpc.sh@64 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -e bdev 00:06:38.867 15:12:28 rpc -- rpc/rpc.sh@66 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:06:38.867 15:12:28 rpc -- rpc/rpc.sh@67 -- # waitforlisten 69008 00:06:38.867 15:12:28 rpc -- common/autotest_common.sh@835 -- # '[' -z 69008 ']' 00:06:38.867 15:12:28 rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:38.867 15:12:28 rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:38.867 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:38.867 15:12:28 rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:38.867 15:12:28 rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:38.867 15:12:28 rpc -- common/autotest_common.sh@10 -- # set +x 00:06:38.867 [2024-11-19 15:12:29.062713] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:06:38.867 [2024-11-19 15:12:29.062866] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69008 ] 00:06:39.127 [2024-11-19 15:12:29.220834] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:39.127 [2024-11-19 15:12:29.260528] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask bdev specified. 00:06:39.127 [2024-11-19 15:12:29.260601] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s spdk_tgt -p 69008' to capture a snapshot of events at runtime. 00:06:39.127 [2024-11-19 15:12:29.260618] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:06:39.127 [2024-11-19 15:12:29.260628] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:06:39.127 [2024-11-19 15:12:29.260648] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/spdk_tgt_trace.pid69008 for offline analysis/debug. 00:06:39.127 [2024-11-19 15:12:29.261143] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:39.696 15:12:29 rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:39.696 15:12:29 rpc -- common/autotest_common.sh@868 -- # return 0 00:06:39.696 15:12:29 rpc -- rpc/rpc.sh@69 -- # export PYTHONPATH=:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/test/rpc 00:06:39.697 15:12:29 rpc -- rpc/rpc.sh@69 -- # PYTHONPATH=:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/test/rpc 00:06:39.697 15:12:29 rpc -- rpc/rpc.sh@72 -- # rpc=rpc_cmd 00:06:39.697 15:12:29 rpc -- rpc/rpc.sh@73 -- # run_test rpc_integrity rpc_integrity 00:06:39.697 15:12:29 rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:39.697 15:12:29 rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:39.697 15:12:29 rpc -- common/autotest_common.sh@10 -- # set +x 00:06:39.697 ************************************ 00:06:39.697 START TEST rpc_integrity 00:06:39.697 ************************************ 00:06:39.697 15:12:29 rpc.rpc_integrity -- common/autotest_common.sh@1129 -- # rpc_integrity 00:06:39.697 15:12:29 rpc.rpc_integrity -- rpc/rpc.sh@12 -- # rpc_cmd bdev_get_bdevs 00:06:39.697 15:12:29 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:39.697 15:12:29 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:39.697 15:12:29 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:39.697 15:12:29 rpc.rpc_integrity -- rpc/rpc.sh@12 -- # bdevs='[]' 00:06:39.697 15:12:29 rpc.rpc_integrity -- rpc/rpc.sh@13 -- # jq length 00:06:39.697 15:12:29 rpc.rpc_integrity -- rpc/rpc.sh@13 -- # '[' 0 == 0 ']' 00:06:39.697 15:12:29 rpc.rpc_integrity -- rpc/rpc.sh@15 -- # rpc_cmd bdev_malloc_create 8 512 00:06:39.697 15:12:29 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:39.697 15:12:29 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:39.697 15:12:29 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:39.697 15:12:29 rpc.rpc_integrity -- rpc/rpc.sh@15 -- # malloc=Malloc0 00:06:39.697 15:12:29 rpc.rpc_integrity -- rpc/rpc.sh@16 -- # rpc_cmd bdev_get_bdevs 00:06:39.697 15:12:29 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:39.697 15:12:29 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:39.697 15:12:29 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:39.697 15:12:29 rpc.rpc_integrity -- rpc/rpc.sh@16 -- # bdevs='[ 00:06:39.697 { 00:06:39.697 "name": "Malloc0", 00:06:39.697 "aliases": [ 00:06:39.697 "fce4ac65-4542-42ff-afbe-7cd334bce0ac" 00:06:39.697 ], 00:06:39.697 "product_name": "Malloc disk", 00:06:39.697 "block_size": 512, 00:06:39.697 "num_blocks": 16384, 00:06:39.697 "uuid": "fce4ac65-4542-42ff-afbe-7cd334bce0ac", 00:06:39.697 "assigned_rate_limits": { 00:06:39.697 "rw_ios_per_sec": 0, 00:06:39.697 "rw_mbytes_per_sec": 0, 00:06:39.697 "r_mbytes_per_sec": 0, 00:06:39.697 "w_mbytes_per_sec": 0 00:06:39.697 }, 00:06:39.697 "claimed": false, 00:06:39.697 "zoned": false, 00:06:39.697 "supported_io_types": { 00:06:39.697 "read": true, 00:06:39.697 "write": true, 00:06:39.697 "unmap": true, 00:06:39.697 "flush": true, 00:06:39.697 "reset": true, 00:06:39.697 "nvme_admin": false, 00:06:39.697 "nvme_io": false, 00:06:39.697 "nvme_io_md": false, 00:06:39.697 "write_zeroes": true, 00:06:39.697 "zcopy": true, 00:06:39.697 "get_zone_info": false, 00:06:39.697 "zone_management": false, 00:06:39.697 "zone_append": false, 00:06:39.697 "compare": false, 00:06:39.697 "compare_and_write": false, 00:06:39.697 "abort": true, 00:06:39.697 "seek_hole": false, 00:06:39.697 "seek_data": false, 00:06:39.697 "copy": true, 00:06:39.697 "nvme_iov_md": false 00:06:39.697 }, 00:06:39.697 "memory_domains": [ 00:06:39.697 { 00:06:39.697 "dma_device_id": "system", 00:06:39.697 "dma_device_type": 1 00:06:39.697 }, 00:06:39.697 { 00:06:39.697 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:39.697 "dma_device_type": 2 00:06:39.697 } 00:06:39.697 ], 00:06:39.697 "driver_specific": {} 00:06:39.697 } 00:06:39.697 ]' 00:06:39.697 15:12:29 rpc.rpc_integrity -- rpc/rpc.sh@17 -- # jq length 00:06:39.958 15:12:30 rpc.rpc_integrity -- rpc/rpc.sh@17 -- # '[' 1 == 1 ']' 00:06:39.958 15:12:30 rpc.rpc_integrity -- rpc/rpc.sh@19 -- # rpc_cmd bdev_passthru_create -b Malloc0 -p Passthru0 00:06:39.958 15:12:30 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:39.958 15:12:30 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:39.958 [2024-11-19 15:12:30.044393] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on Malloc0 00:06:39.958 [2024-11-19 15:12:30.044478] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:06:39.958 [2024-11-19 15:12:30.044521] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006c80 00:06:39.958 [2024-11-19 15:12:30.044532] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:06:39.958 [2024-11-19 15:12:30.047328] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:06:39.958 [2024-11-19 15:12:30.047368] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: Passthru0 00:06:39.958 Passthru0 00:06:39.958 15:12:30 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:39.958 15:12:30 rpc.rpc_integrity -- rpc/rpc.sh@20 -- # rpc_cmd bdev_get_bdevs 00:06:39.958 15:12:30 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:39.958 15:12:30 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:39.958 15:12:30 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:39.958 15:12:30 rpc.rpc_integrity -- rpc/rpc.sh@20 -- # bdevs='[ 00:06:39.958 { 00:06:39.958 "name": "Malloc0", 00:06:39.958 "aliases": [ 00:06:39.958 "fce4ac65-4542-42ff-afbe-7cd334bce0ac" 00:06:39.958 ], 00:06:39.958 "product_name": "Malloc disk", 00:06:39.958 "block_size": 512, 00:06:39.958 "num_blocks": 16384, 00:06:39.958 "uuid": "fce4ac65-4542-42ff-afbe-7cd334bce0ac", 00:06:39.958 "assigned_rate_limits": { 00:06:39.958 "rw_ios_per_sec": 0, 00:06:39.958 "rw_mbytes_per_sec": 0, 00:06:39.958 "r_mbytes_per_sec": 0, 00:06:39.958 "w_mbytes_per_sec": 0 00:06:39.958 }, 00:06:39.958 "claimed": true, 00:06:39.958 "claim_type": "exclusive_write", 00:06:39.958 "zoned": false, 00:06:39.958 "supported_io_types": { 00:06:39.958 "read": true, 00:06:39.958 "write": true, 00:06:39.958 "unmap": true, 00:06:39.958 "flush": true, 00:06:39.958 "reset": true, 00:06:39.958 "nvme_admin": false, 00:06:39.958 "nvme_io": false, 00:06:39.958 "nvme_io_md": false, 00:06:39.958 "write_zeroes": true, 00:06:39.958 "zcopy": true, 00:06:39.958 "get_zone_info": false, 00:06:39.958 "zone_management": false, 00:06:39.958 "zone_append": false, 00:06:39.958 "compare": false, 00:06:39.958 "compare_and_write": false, 00:06:39.958 "abort": true, 00:06:39.958 "seek_hole": false, 00:06:39.958 "seek_data": false, 00:06:39.958 "copy": true, 00:06:39.958 "nvme_iov_md": false 00:06:39.958 }, 00:06:39.958 "memory_domains": [ 00:06:39.958 { 00:06:39.958 "dma_device_id": "system", 00:06:39.958 "dma_device_type": 1 00:06:39.958 }, 00:06:39.958 { 00:06:39.958 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:39.958 "dma_device_type": 2 00:06:39.958 } 00:06:39.958 ], 00:06:39.958 "driver_specific": {} 00:06:39.958 }, 00:06:39.958 { 00:06:39.958 "name": "Passthru0", 00:06:39.958 "aliases": [ 00:06:39.958 "605ae048-b4d5-5065-a87f-3a77acae1b8b" 00:06:39.958 ], 00:06:39.958 "product_name": "passthru", 00:06:39.958 "block_size": 512, 00:06:39.958 "num_blocks": 16384, 00:06:39.958 "uuid": "605ae048-b4d5-5065-a87f-3a77acae1b8b", 00:06:39.958 "assigned_rate_limits": { 00:06:39.958 "rw_ios_per_sec": 0, 00:06:39.958 "rw_mbytes_per_sec": 0, 00:06:39.958 "r_mbytes_per_sec": 0, 00:06:39.958 "w_mbytes_per_sec": 0 00:06:39.958 }, 00:06:39.958 "claimed": false, 00:06:39.958 "zoned": false, 00:06:39.958 "supported_io_types": { 00:06:39.958 "read": true, 00:06:39.958 "write": true, 00:06:39.958 "unmap": true, 00:06:39.958 "flush": true, 00:06:39.958 "reset": true, 00:06:39.958 "nvme_admin": false, 00:06:39.958 "nvme_io": false, 00:06:39.958 "nvme_io_md": false, 00:06:39.958 "write_zeroes": true, 00:06:39.958 "zcopy": true, 00:06:39.958 "get_zone_info": false, 00:06:39.958 "zone_management": false, 00:06:39.958 "zone_append": false, 00:06:39.958 "compare": false, 00:06:39.958 "compare_and_write": false, 00:06:39.958 "abort": true, 00:06:39.958 "seek_hole": false, 00:06:39.958 "seek_data": false, 00:06:39.958 "copy": true, 00:06:39.958 "nvme_iov_md": false 00:06:39.958 }, 00:06:39.958 "memory_domains": [ 00:06:39.958 { 00:06:39.958 "dma_device_id": "system", 00:06:39.958 "dma_device_type": 1 00:06:39.958 }, 00:06:39.958 { 00:06:39.958 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:39.958 "dma_device_type": 2 00:06:39.958 } 00:06:39.958 ], 00:06:39.958 "driver_specific": { 00:06:39.958 "passthru": { 00:06:39.958 "name": "Passthru0", 00:06:39.958 "base_bdev_name": "Malloc0" 00:06:39.958 } 00:06:39.958 } 00:06:39.958 } 00:06:39.958 ]' 00:06:39.958 15:12:30 rpc.rpc_integrity -- rpc/rpc.sh@21 -- # jq length 00:06:39.958 15:12:30 rpc.rpc_integrity -- rpc/rpc.sh@21 -- # '[' 2 == 2 ']' 00:06:39.958 15:12:30 rpc.rpc_integrity -- rpc/rpc.sh@23 -- # rpc_cmd bdev_passthru_delete Passthru0 00:06:39.958 15:12:30 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:39.958 15:12:30 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:39.958 15:12:30 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:39.958 15:12:30 rpc.rpc_integrity -- rpc/rpc.sh@24 -- # rpc_cmd bdev_malloc_delete Malloc0 00:06:39.958 15:12:30 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:39.958 15:12:30 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:39.958 15:12:30 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:39.958 15:12:30 rpc.rpc_integrity -- rpc/rpc.sh@25 -- # rpc_cmd bdev_get_bdevs 00:06:39.958 15:12:30 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:39.958 15:12:30 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:39.958 15:12:30 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:39.958 15:12:30 rpc.rpc_integrity -- rpc/rpc.sh@25 -- # bdevs='[]' 00:06:39.958 15:12:30 rpc.rpc_integrity -- rpc/rpc.sh@26 -- # jq length 00:06:39.958 15:12:30 rpc.rpc_integrity -- rpc/rpc.sh@26 -- # '[' 0 == 0 ']' 00:06:39.958 00:06:39.958 real 0m0.321s 00:06:39.958 user 0m0.185s 00:06:39.958 sys 0m0.058s 00:06:39.959 15:12:30 rpc.rpc_integrity -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:39.959 15:12:30 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:39.959 ************************************ 00:06:39.959 END TEST rpc_integrity 00:06:39.959 ************************************ 00:06:39.959 15:12:30 rpc -- rpc/rpc.sh@74 -- # run_test rpc_plugins rpc_plugins 00:06:39.959 15:12:30 rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:39.959 15:12:30 rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:39.959 15:12:30 rpc -- common/autotest_common.sh@10 -- # set +x 00:06:39.959 ************************************ 00:06:39.959 START TEST rpc_plugins 00:06:39.959 ************************************ 00:06:39.959 15:12:30 rpc.rpc_plugins -- common/autotest_common.sh@1129 -- # rpc_plugins 00:06:39.959 15:12:30 rpc.rpc_plugins -- rpc/rpc.sh@30 -- # rpc_cmd --plugin rpc_plugin create_malloc 00:06:39.959 15:12:30 rpc.rpc_plugins -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:39.959 15:12:30 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:06:40.219 15:12:30 rpc.rpc_plugins -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:40.219 15:12:30 rpc.rpc_plugins -- rpc/rpc.sh@30 -- # malloc=Malloc1 00:06:40.219 15:12:30 rpc.rpc_plugins -- rpc/rpc.sh@31 -- # rpc_cmd bdev_get_bdevs 00:06:40.219 15:12:30 rpc.rpc_plugins -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:40.219 15:12:30 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:06:40.219 15:12:30 rpc.rpc_plugins -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:40.219 15:12:30 rpc.rpc_plugins -- rpc/rpc.sh@31 -- # bdevs='[ 00:06:40.219 { 00:06:40.219 "name": "Malloc1", 00:06:40.219 "aliases": [ 00:06:40.219 "49fb874d-4f79-4387-9bce-cbf981b321fa" 00:06:40.219 ], 00:06:40.219 "product_name": "Malloc disk", 00:06:40.219 "block_size": 4096, 00:06:40.219 "num_blocks": 256, 00:06:40.219 "uuid": "49fb874d-4f79-4387-9bce-cbf981b321fa", 00:06:40.219 "assigned_rate_limits": { 00:06:40.219 "rw_ios_per_sec": 0, 00:06:40.219 "rw_mbytes_per_sec": 0, 00:06:40.219 "r_mbytes_per_sec": 0, 00:06:40.219 "w_mbytes_per_sec": 0 00:06:40.219 }, 00:06:40.219 "claimed": false, 00:06:40.219 "zoned": false, 00:06:40.219 "supported_io_types": { 00:06:40.219 "read": true, 00:06:40.219 "write": true, 00:06:40.219 "unmap": true, 00:06:40.219 "flush": true, 00:06:40.219 "reset": true, 00:06:40.219 "nvme_admin": false, 00:06:40.219 "nvme_io": false, 00:06:40.219 "nvme_io_md": false, 00:06:40.219 "write_zeroes": true, 00:06:40.219 "zcopy": true, 00:06:40.219 "get_zone_info": false, 00:06:40.219 "zone_management": false, 00:06:40.219 "zone_append": false, 00:06:40.219 "compare": false, 00:06:40.219 "compare_and_write": false, 00:06:40.219 "abort": true, 00:06:40.219 "seek_hole": false, 00:06:40.219 "seek_data": false, 00:06:40.219 "copy": true, 00:06:40.219 "nvme_iov_md": false 00:06:40.219 }, 00:06:40.219 "memory_domains": [ 00:06:40.219 { 00:06:40.219 "dma_device_id": "system", 00:06:40.219 "dma_device_type": 1 00:06:40.219 }, 00:06:40.219 { 00:06:40.219 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:40.219 "dma_device_type": 2 00:06:40.219 } 00:06:40.219 ], 00:06:40.219 "driver_specific": {} 00:06:40.219 } 00:06:40.219 ]' 00:06:40.219 15:12:30 rpc.rpc_plugins -- rpc/rpc.sh@32 -- # jq length 00:06:40.219 15:12:30 rpc.rpc_plugins -- rpc/rpc.sh@32 -- # '[' 1 == 1 ']' 00:06:40.219 15:12:30 rpc.rpc_plugins -- rpc/rpc.sh@34 -- # rpc_cmd --plugin rpc_plugin delete_malloc Malloc1 00:06:40.219 15:12:30 rpc.rpc_plugins -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:40.219 15:12:30 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:06:40.219 15:12:30 rpc.rpc_plugins -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:40.219 15:12:30 rpc.rpc_plugins -- rpc/rpc.sh@35 -- # rpc_cmd bdev_get_bdevs 00:06:40.219 15:12:30 rpc.rpc_plugins -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:40.219 15:12:30 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:06:40.219 15:12:30 rpc.rpc_plugins -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:40.219 15:12:30 rpc.rpc_plugins -- rpc/rpc.sh@35 -- # bdevs='[]' 00:06:40.219 15:12:30 rpc.rpc_plugins -- rpc/rpc.sh@36 -- # jq length 00:06:40.219 15:12:30 rpc.rpc_plugins -- rpc/rpc.sh@36 -- # '[' 0 == 0 ']' 00:06:40.219 00:06:40.219 real 0m0.155s 00:06:40.219 user 0m0.088s 00:06:40.219 sys 0m0.024s 00:06:40.219 15:12:30 rpc.rpc_plugins -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:40.219 15:12:30 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:06:40.219 ************************************ 00:06:40.219 END TEST rpc_plugins 00:06:40.219 ************************************ 00:06:40.219 15:12:30 rpc -- rpc/rpc.sh@75 -- # run_test rpc_trace_cmd_test rpc_trace_cmd_test 00:06:40.219 15:12:30 rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:40.219 15:12:30 rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:40.219 15:12:30 rpc -- common/autotest_common.sh@10 -- # set +x 00:06:40.219 ************************************ 00:06:40.219 START TEST rpc_trace_cmd_test 00:06:40.219 ************************************ 00:06:40.219 15:12:30 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@1129 -- # rpc_trace_cmd_test 00:06:40.219 15:12:30 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@40 -- # local info 00:06:40.219 15:12:30 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@42 -- # rpc_cmd trace_get_info 00:06:40.219 15:12:30 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:40.219 15:12:30 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@10 -- # set +x 00:06:40.219 15:12:30 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:40.219 15:12:30 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@42 -- # info='{ 00:06:40.219 "tpoint_shm_path": "/dev/shm/spdk_tgt_trace.pid69008", 00:06:40.219 "tpoint_group_mask": "0x8", 00:06:40.219 "iscsi_conn": { 00:06:40.219 "mask": "0x2", 00:06:40.219 "tpoint_mask": "0x0" 00:06:40.219 }, 00:06:40.219 "scsi": { 00:06:40.219 "mask": "0x4", 00:06:40.219 "tpoint_mask": "0x0" 00:06:40.219 }, 00:06:40.219 "bdev": { 00:06:40.219 "mask": "0x8", 00:06:40.219 "tpoint_mask": "0xffffffffffffffff" 00:06:40.219 }, 00:06:40.219 "nvmf_rdma": { 00:06:40.219 "mask": "0x10", 00:06:40.219 "tpoint_mask": "0x0" 00:06:40.219 }, 00:06:40.219 "nvmf_tcp": { 00:06:40.219 "mask": "0x20", 00:06:40.219 "tpoint_mask": "0x0" 00:06:40.219 }, 00:06:40.219 "ftl": { 00:06:40.219 "mask": "0x40", 00:06:40.219 "tpoint_mask": "0x0" 00:06:40.219 }, 00:06:40.219 "blobfs": { 00:06:40.219 "mask": "0x80", 00:06:40.219 "tpoint_mask": "0x0" 00:06:40.219 }, 00:06:40.219 "dsa": { 00:06:40.219 "mask": "0x200", 00:06:40.219 "tpoint_mask": "0x0" 00:06:40.219 }, 00:06:40.219 "thread": { 00:06:40.219 "mask": "0x400", 00:06:40.219 "tpoint_mask": "0x0" 00:06:40.219 }, 00:06:40.219 "nvme_pcie": { 00:06:40.219 "mask": "0x800", 00:06:40.219 "tpoint_mask": "0x0" 00:06:40.219 }, 00:06:40.219 "iaa": { 00:06:40.219 "mask": "0x1000", 00:06:40.219 "tpoint_mask": "0x0" 00:06:40.219 }, 00:06:40.219 "nvme_tcp": { 00:06:40.219 "mask": "0x2000", 00:06:40.219 "tpoint_mask": "0x0" 00:06:40.219 }, 00:06:40.219 "bdev_nvme": { 00:06:40.219 "mask": "0x4000", 00:06:40.219 "tpoint_mask": "0x0" 00:06:40.219 }, 00:06:40.219 "sock": { 00:06:40.219 "mask": "0x8000", 00:06:40.220 "tpoint_mask": "0x0" 00:06:40.220 }, 00:06:40.220 "blob": { 00:06:40.220 "mask": "0x10000", 00:06:40.220 "tpoint_mask": "0x0" 00:06:40.220 }, 00:06:40.220 "bdev_raid": { 00:06:40.220 "mask": "0x20000", 00:06:40.220 "tpoint_mask": "0x0" 00:06:40.220 }, 00:06:40.220 "scheduler": { 00:06:40.220 "mask": "0x40000", 00:06:40.220 "tpoint_mask": "0x0" 00:06:40.220 } 00:06:40.220 }' 00:06:40.220 15:12:30 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@43 -- # jq length 00:06:40.480 15:12:30 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@43 -- # '[' 19 -gt 2 ']' 00:06:40.480 15:12:30 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@44 -- # jq 'has("tpoint_group_mask")' 00:06:40.480 15:12:30 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@44 -- # '[' true = true ']' 00:06:40.480 15:12:30 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@45 -- # jq 'has("tpoint_shm_path")' 00:06:40.480 15:12:30 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@45 -- # '[' true = true ']' 00:06:40.480 15:12:30 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@46 -- # jq 'has("bdev")' 00:06:40.480 15:12:30 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@46 -- # '[' true = true ']' 00:06:40.480 15:12:30 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@47 -- # jq -r .bdev.tpoint_mask 00:06:40.480 15:12:30 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@47 -- # '[' 0xffffffffffffffff '!=' 0x0 ']' 00:06:40.480 00:06:40.480 real 0m0.247s 00:06:40.480 user 0m0.207s 00:06:40.480 sys 0m0.032s 00:06:40.480 15:12:30 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:40.480 15:12:30 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@10 -- # set +x 00:06:40.480 ************************************ 00:06:40.480 END TEST rpc_trace_cmd_test 00:06:40.480 ************************************ 00:06:40.480 15:12:30 rpc -- rpc/rpc.sh@76 -- # [[ 0 -eq 1 ]] 00:06:40.480 15:12:30 rpc -- rpc/rpc.sh@80 -- # rpc=rpc_cmd 00:06:40.480 15:12:30 rpc -- rpc/rpc.sh@81 -- # run_test rpc_daemon_integrity rpc_integrity 00:06:40.480 15:12:30 rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:40.480 15:12:30 rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:40.480 15:12:30 rpc -- common/autotest_common.sh@10 -- # set +x 00:06:40.480 ************************************ 00:06:40.480 START TEST rpc_daemon_integrity 00:06:40.480 ************************************ 00:06:40.480 15:12:30 rpc.rpc_daemon_integrity -- common/autotest_common.sh@1129 -- # rpc_integrity 00:06:40.741 15:12:30 rpc.rpc_daemon_integrity -- rpc/rpc.sh@12 -- # rpc_cmd bdev_get_bdevs 00:06:40.741 15:12:30 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:40.741 15:12:30 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:40.741 15:12:30 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:40.741 15:12:30 rpc.rpc_daemon_integrity -- rpc/rpc.sh@12 -- # bdevs='[]' 00:06:40.741 15:12:30 rpc.rpc_daemon_integrity -- rpc/rpc.sh@13 -- # jq length 00:06:40.741 15:12:30 rpc.rpc_daemon_integrity -- rpc/rpc.sh@13 -- # '[' 0 == 0 ']' 00:06:40.741 15:12:30 rpc.rpc_daemon_integrity -- rpc/rpc.sh@15 -- # rpc_cmd bdev_malloc_create 8 512 00:06:40.741 15:12:30 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:40.741 15:12:30 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:40.741 15:12:30 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:40.741 15:12:30 rpc.rpc_daemon_integrity -- rpc/rpc.sh@15 -- # malloc=Malloc2 00:06:40.741 15:12:30 rpc.rpc_daemon_integrity -- rpc/rpc.sh@16 -- # rpc_cmd bdev_get_bdevs 00:06:40.741 15:12:30 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:40.741 15:12:30 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:40.741 15:12:30 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:40.741 15:12:30 rpc.rpc_daemon_integrity -- rpc/rpc.sh@16 -- # bdevs='[ 00:06:40.741 { 00:06:40.741 "name": "Malloc2", 00:06:40.741 "aliases": [ 00:06:40.741 "a0ae0a6a-4704-4f37-8f34-5e419ee320ed" 00:06:40.741 ], 00:06:40.741 "product_name": "Malloc disk", 00:06:40.741 "block_size": 512, 00:06:40.741 "num_blocks": 16384, 00:06:40.741 "uuid": "a0ae0a6a-4704-4f37-8f34-5e419ee320ed", 00:06:40.741 "assigned_rate_limits": { 00:06:40.741 "rw_ios_per_sec": 0, 00:06:40.741 "rw_mbytes_per_sec": 0, 00:06:40.741 "r_mbytes_per_sec": 0, 00:06:40.741 "w_mbytes_per_sec": 0 00:06:40.741 }, 00:06:40.741 "claimed": false, 00:06:40.741 "zoned": false, 00:06:40.741 "supported_io_types": { 00:06:40.741 "read": true, 00:06:40.741 "write": true, 00:06:40.741 "unmap": true, 00:06:40.741 "flush": true, 00:06:40.741 "reset": true, 00:06:40.741 "nvme_admin": false, 00:06:40.741 "nvme_io": false, 00:06:40.741 "nvme_io_md": false, 00:06:40.741 "write_zeroes": true, 00:06:40.741 "zcopy": true, 00:06:40.741 "get_zone_info": false, 00:06:40.741 "zone_management": false, 00:06:40.741 "zone_append": false, 00:06:40.741 "compare": false, 00:06:40.741 "compare_and_write": false, 00:06:40.741 "abort": true, 00:06:40.741 "seek_hole": false, 00:06:40.741 "seek_data": false, 00:06:40.741 "copy": true, 00:06:40.741 "nvme_iov_md": false 00:06:40.741 }, 00:06:40.741 "memory_domains": [ 00:06:40.741 { 00:06:40.741 "dma_device_id": "system", 00:06:40.741 "dma_device_type": 1 00:06:40.741 }, 00:06:40.741 { 00:06:40.741 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:40.741 "dma_device_type": 2 00:06:40.741 } 00:06:40.741 ], 00:06:40.741 "driver_specific": {} 00:06:40.741 } 00:06:40.741 ]' 00:06:40.741 15:12:30 rpc.rpc_daemon_integrity -- rpc/rpc.sh@17 -- # jq length 00:06:40.741 15:12:30 rpc.rpc_daemon_integrity -- rpc/rpc.sh@17 -- # '[' 1 == 1 ']' 00:06:40.741 15:12:30 rpc.rpc_daemon_integrity -- rpc/rpc.sh@19 -- # rpc_cmd bdev_passthru_create -b Malloc2 -p Passthru0 00:06:40.741 15:12:30 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:40.741 15:12:30 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:40.741 [2024-11-19 15:12:30.974581] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on Malloc2 00:06:40.741 [2024-11-19 15:12:30.974653] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:06:40.741 [2024-11-19 15:12:30.974680] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:06:40.742 [2024-11-19 15:12:30.974690] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:06:40.742 [2024-11-19 15:12:30.977307] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:06:40.742 [2024-11-19 15:12:30.977343] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: Passthru0 00:06:40.742 Passthru0 00:06:40.742 15:12:30 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:40.742 15:12:30 rpc.rpc_daemon_integrity -- rpc/rpc.sh@20 -- # rpc_cmd bdev_get_bdevs 00:06:40.742 15:12:30 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:40.742 15:12:30 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:40.742 15:12:31 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:40.742 15:12:31 rpc.rpc_daemon_integrity -- rpc/rpc.sh@20 -- # bdevs='[ 00:06:40.742 { 00:06:40.742 "name": "Malloc2", 00:06:40.742 "aliases": [ 00:06:40.742 "a0ae0a6a-4704-4f37-8f34-5e419ee320ed" 00:06:40.742 ], 00:06:40.742 "product_name": "Malloc disk", 00:06:40.742 "block_size": 512, 00:06:40.742 "num_blocks": 16384, 00:06:40.742 "uuid": "a0ae0a6a-4704-4f37-8f34-5e419ee320ed", 00:06:40.742 "assigned_rate_limits": { 00:06:40.742 "rw_ios_per_sec": 0, 00:06:40.742 "rw_mbytes_per_sec": 0, 00:06:40.742 "r_mbytes_per_sec": 0, 00:06:40.742 "w_mbytes_per_sec": 0 00:06:40.742 }, 00:06:40.742 "claimed": true, 00:06:40.742 "claim_type": "exclusive_write", 00:06:40.742 "zoned": false, 00:06:40.742 "supported_io_types": { 00:06:40.742 "read": true, 00:06:40.742 "write": true, 00:06:40.742 "unmap": true, 00:06:40.742 "flush": true, 00:06:40.742 "reset": true, 00:06:40.742 "nvme_admin": false, 00:06:40.742 "nvme_io": false, 00:06:40.742 "nvme_io_md": false, 00:06:40.742 "write_zeroes": true, 00:06:40.742 "zcopy": true, 00:06:40.742 "get_zone_info": false, 00:06:40.742 "zone_management": false, 00:06:40.742 "zone_append": false, 00:06:40.742 "compare": false, 00:06:40.742 "compare_and_write": false, 00:06:40.742 "abort": true, 00:06:40.742 "seek_hole": false, 00:06:40.742 "seek_data": false, 00:06:40.742 "copy": true, 00:06:40.742 "nvme_iov_md": false 00:06:40.742 }, 00:06:40.742 "memory_domains": [ 00:06:40.742 { 00:06:40.742 "dma_device_id": "system", 00:06:40.742 "dma_device_type": 1 00:06:40.742 }, 00:06:40.742 { 00:06:40.742 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:40.742 "dma_device_type": 2 00:06:40.742 } 00:06:40.742 ], 00:06:40.742 "driver_specific": {} 00:06:40.742 }, 00:06:40.742 { 00:06:40.742 "name": "Passthru0", 00:06:40.742 "aliases": [ 00:06:40.742 "735f90a8-ee85-55dc-8dfc-8f0c68d3a53d" 00:06:40.742 ], 00:06:40.742 "product_name": "passthru", 00:06:40.742 "block_size": 512, 00:06:40.742 "num_blocks": 16384, 00:06:40.742 "uuid": "735f90a8-ee85-55dc-8dfc-8f0c68d3a53d", 00:06:40.742 "assigned_rate_limits": { 00:06:40.742 "rw_ios_per_sec": 0, 00:06:40.742 "rw_mbytes_per_sec": 0, 00:06:40.742 "r_mbytes_per_sec": 0, 00:06:40.742 "w_mbytes_per_sec": 0 00:06:40.742 }, 00:06:40.742 "claimed": false, 00:06:40.742 "zoned": false, 00:06:40.742 "supported_io_types": { 00:06:40.742 "read": true, 00:06:40.742 "write": true, 00:06:40.742 "unmap": true, 00:06:40.742 "flush": true, 00:06:40.742 "reset": true, 00:06:40.742 "nvme_admin": false, 00:06:40.742 "nvme_io": false, 00:06:40.742 "nvme_io_md": false, 00:06:40.742 "write_zeroes": true, 00:06:40.742 "zcopy": true, 00:06:40.742 "get_zone_info": false, 00:06:40.742 "zone_management": false, 00:06:40.742 "zone_append": false, 00:06:40.742 "compare": false, 00:06:40.742 "compare_and_write": false, 00:06:40.742 "abort": true, 00:06:40.742 "seek_hole": false, 00:06:40.742 "seek_data": false, 00:06:40.742 "copy": true, 00:06:40.742 "nvme_iov_md": false 00:06:40.742 }, 00:06:40.742 "memory_domains": [ 00:06:40.742 { 00:06:40.742 "dma_device_id": "system", 00:06:40.742 "dma_device_type": 1 00:06:40.742 }, 00:06:40.742 { 00:06:40.742 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:40.742 "dma_device_type": 2 00:06:40.742 } 00:06:40.742 ], 00:06:40.742 "driver_specific": { 00:06:40.742 "passthru": { 00:06:40.742 "name": "Passthru0", 00:06:40.742 "base_bdev_name": "Malloc2" 00:06:40.742 } 00:06:40.742 } 00:06:40.742 } 00:06:40.742 ]' 00:06:40.742 15:12:31 rpc.rpc_daemon_integrity -- rpc/rpc.sh@21 -- # jq length 00:06:40.742 15:12:31 rpc.rpc_daemon_integrity -- rpc/rpc.sh@21 -- # '[' 2 == 2 ']' 00:06:40.742 15:12:31 rpc.rpc_daemon_integrity -- rpc/rpc.sh@23 -- # rpc_cmd bdev_passthru_delete Passthru0 00:06:40.742 15:12:31 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:40.742 15:12:31 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:40.742 15:12:31 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:40.742 15:12:31 rpc.rpc_daemon_integrity -- rpc/rpc.sh@24 -- # rpc_cmd bdev_malloc_delete Malloc2 00:06:40.742 15:12:31 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:40.742 15:12:31 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:41.003 15:12:31 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:41.003 15:12:31 rpc.rpc_daemon_integrity -- rpc/rpc.sh@25 -- # rpc_cmd bdev_get_bdevs 00:06:41.003 15:12:31 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:41.003 15:12:31 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:41.003 15:12:31 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:41.003 15:12:31 rpc.rpc_daemon_integrity -- rpc/rpc.sh@25 -- # bdevs='[]' 00:06:41.003 15:12:31 rpc.rpc_daemon_integrity -- rpc/rpc.sh@26 -- # jq length 00:06:41.003 15:12:31 rpc.rpc_daemon_integrity -- rpc/rpc.sh@26 -- # '[' 0 == 0 ']' 00:06:41.003 00:06:41.003 real 0m0.321s 00:06:41.003 user 0m0.188s 00:06:41.003 sys 0m0.058s 00:06:41.003 15:12:31 rpc.rpc_daemon_integrity -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:41.003 15:12:31 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:41.003 ************************************ 00:06:41.003 END TEST rpc_daemon_integrity 00:06:41.003 ************************************ 00:06:41.003 15:12:31 rpc -- rpc/rpc.sh@83 -- # trap - SIGINT SIGTERM EXIT 00:06:41.003 15:12:31 rpc -- rpc/rpc.sh@84 -- # killprocess 69008 00:06:41.003 15:12:31 rpc -- common/autotest_common.sh@954 -- # '[' -z 69008 ']' 00:06:41.003 15:12:31 rpc -- common/autotest_common.sh@958 -- # kill -0 69008 00:06:41.003 15:12:31 rpc -- common/autotest_common.sh@959 -- # uname 00:06:41.003 15:12:31 rpc -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:41.003 15:12:31 rpc -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 69008 00:06:41.003 15:12:31 rpc -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:41.003 15:12:31 rpc -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:41.003 killing process with pid 69008 00:06:41.003 15:12:31 rpc -- common/autotest_common.sh@972 -- # echo 'killing process with pid 69008' 00:06:41.003 15:12:31 rpc -- common/autotest_common.sh@973 -- # kill 69008 00:06:41.003 15:12:31 rpc -- common/autotest_common.sh@978 -- # wait 69008 00:06:41.583 00:06:41.583 real 0m3.104s 00:06:41.583 user 0m3.546s 00:06:41.583 sys 0m1.016s 00:06:41.583 15:12:31 rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:41.583 15:12:31 rpc -- common/autotest_common.sh@10 -- # set +x 00:06:41.583 ************************************ 00:06:41.583 END TEST rpc 00:06:41.583 ************************************ 00:06:41.583 15:12:31 -- spdk/autotest.sh@157 -- # run_test skip_rpc /home/vagrant/spdk_repo/spdk/test/rpc/skip_rpc.sh 00:06:41.583 15:12:31 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:41.583 15:12:31 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:41.583 15:12:31 -- common/autotest_common.sh@10 -- # set +x 00:06:41.873 ************************************ 00:06:41.873 START TEST skip_rpc 00:06:41.873 ************************************ 00:06:41.873 15:12:31 skip_rpc -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/rpc/skip_rpc.sh 00:06:41.873 * Looking for test storage... 00:06:41.873 * Found test storage at /home/vagrant/spdk_repo/spdk/test/rpc 00:06:41.873 15:12:32 skip_rpc -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:06:41.873 15:12:32 skip_rpc -- common/autotest_common.sh@1693 -- # lcov --version 00:06:41.873 15:12:32 skip_rpc -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:06:41.873 15:12:32 skip_rpc -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:06:41.874 15:12:32 skip_rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:41.874 15:12:32 skip_rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:41.874 15:12:32 skip_rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:41.874 15:12:32 skip_rpc -- scripts/common.sh@336 -- # IFS=.-: 00:06:41.874 15:12:32 skip_rpc -- scripts/common.sh@336 -- # read -ra ver1 00:06:41.874 15:12:32 skip_rpc -- scripts/common.sh@337 -- # IFS=.-: 00:06:41.874 15:12:32 skip_rpc -- scripts/common.sh@337 -- # read -ra ver2 00:06:41.874 15:12:32 skip_rpc -- scripts/common.sh@338 -- # local 'op=<' 00:06:41.874 15:12:32 skip_rpc -- scripts/common.sh@340 -- # ver1_l=2 00:06:41.874 15:12:32 skip_rpc -- scripts/common.sh@341 -- # ver2_l=1 00:06:41.874 15:12:32 skip_rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:41.874 15:12:32 skip_rpc -- scripts/common.sh@344 -- # case "$op" in 00:06:41.874 15:12:32 skip_rpc -- scripts/common.sh@345 -- # : 1 00:06:41.874 15:12:32 skip_rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:41.874 15:12:32 skip_rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:41.874 15:12:32 skip_rpc -- scripts/common.sh@365 -- # decimal 1 00:06:41.874 15:12:32 skip_rpc -- scripts/common.sh@353 -- # local d=1 00:06:41.874 15:12:32 skip_rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:41.874 15:12:32 skip_rpc -- scripts/common.sh@355 -- # echo 1 00:06:41.874 15:12:32 skip_rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:06:41.874 15:12:32 skip_rpc -- scripts/common.sh@366 -- # decimal 2 00:06:41.874 15:12:32 skip_rpc -- scripts/common.sh@353 -- # local d=2 00:06:41.874 15:12:32 skip_rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:41.874 15:12:32 skip_rpc -- scripts/common.sh@355 -- # echo 2 00:06:41.874 15:12:32 skip_rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:06:41.874 15:12:32 skip_rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:41.874 15:12:32 skip_rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:41.874 15:12:32 skip_rpc -- scripts/common.sh@368 -- # return 0 00:06:41.874 15:12:32 skip_rpc -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:41.874 15:12:32 skip_rpc -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:06:41.874 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:41.874 --rc genhtml_branch_coverage=1 00:06:41.874 --rc genhtml_function_coverage=1 00:06:41.874 --rc genhtml_legend=1 00:06:41.874 --rc geninfo_all_blocks=1 00:06:41.874 --rc geninfo_unexecuted_blocks=1 00:06:41.874 00:06:41.874 ' 00:06:41.874 15:12:32 skip_rpc -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:06:41.874 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:41.874 --rc genhtml_branch_coverage=1 00:06:41.874 --rc genhtml_function_coverage=1 00:06:41.874 --rc genhtml_legend=1 00:06:41.874 --rc geninfo_all_blocks=1 00:06:41.874 --rc geninfo_unexecuted_blocks=1 00:06:41.874 00:06:41.874 ' 00:06:41.874 15:12:32 skip_rpc -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:06:41.874 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:41.874 --rc genhtml_branch_coverage=1 00:06:41.874 --rc genhtml_function_coverage=1 00:06:41.874 --rc genhtml_legend=1 00:06:41.874 --rc geninfo_all_blocks=1 00:06:41.874 --rc geninfo_unexecuted_blocks=1 00:06:41.874 00:06:41.874 ' 00:06:41.874 15:12:32 skip_rpc -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:06:41.874 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:41.874 --rc genhtml_branch_coverage=1 00:06:41.874 --rc genhtml_function_coverage=1 00:06:41.874 --rc genhtml_legend=1 00:06:41.874 --rc geninfo_all_blocks=1 00:06:41.874 --rc geninfo_unexecuted_blocks=1 00:06:41.874 00:06:41.874 ' 00:06:41.874 15:12:32 skip_rpc -- rpc/skip_rpc.sh@11 -- # CONFIG_PATH=/home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:06:41.874 15:12:32 skip_rpc -- rpc/skip_rpc.sh@12 -- # LOG_PATH=/home/vagrant/spdk_repo/spdk/test/rpc/log.txt 00:06:41.874 15:12:32 skip_rpc -- rpc/skip_rpc.sh@73 -- # run_test skip_rpc test_skip_rpc 00:06:41.874 15:12:32 skip_rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:41.874 15:12:32 skip_rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:41.874 15:12:32 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:41.874 ************************************ 00:06:41.874 START TEST skip_rpc 00:06:41.874 ************************************ 00:06:41.874 15:12:32 skip_rpc.skip_rpc -- common/autotest_common.sh@1129 -- # test_skip_rpc 00:06:41.874 15:12:32 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@16 -- # local spdk_pid=69215 00:06:41.874 15:12:32 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@15 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 00:06:41.874 15:12:32 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@18 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:06:41.874 15:12:32 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@19 -- # sleep 5 00:06:42.135 [2024-11-19 15:12:32.260343] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:06:42.135 [2024-11-19 15:12:32.260512] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69215 ] 00:06:42.135 [2024-11-19 15:12:32.416655] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:42.135 [2024-11-19 15:12:32.457903] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:47.417 15:12:37 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@21 -- # NOT rpc_cmd spdk_get_version 00:06:47.417 15:12:37 skip_rpc.skip_rpc -- common/autotest_common.sh@652 -- # local es=0 00:06:47.417 15:12:37 skip_rpc.skip_rpc -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd spdk_get_version 00:06:47.417 15:12:37 skip_rpc.skip_rpc -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:06:47.417 15:12:37 skip_rpc.skip_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:47.417 15:12:37 skip_rpc.skip_rpc -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:06:47.417 15:12:37 skip_rpc.skip_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:47.417 15:12:37 skip_rpc.skip_rpc -- common/autotest_common.sh@655 -- # rpc_cmd spdk_get_version 00:06:47.417 15:12:37 skip_rpc.skip_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:47.417 15:12:37 skip_rpc.skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:47.417 15:12:37 skip_rpc.skip_rpc -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:06:47.417 15:12:37 skip_rpc.skip_rpc -- common/autotest_common.sh@655 -- # es=1 00:06:47.417 15:12:37 skip_rpc.skip_rpc -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:06:47.417 15:12:37 skip_rpc.skip_rpc -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:06:47.417 15:12:37 skip_rpc.skip_rpc -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:06:47.417 15:12:37 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@22 -- # trap - SIGINT SIGTERM EXIT 00:06:47.417 15:12:37 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@23 -- # killprocess 69215 00:06:47.417 15:12:37 skip_rpc.skip_rpc -- common/autotest_common.sh@954 -- # '[' -z 69215 ']' 00:06:47.417 15:12:37 skip_rpc.skip_rpc -- common/autotest_common.sh@958 -- # kill -0 69215 00:06:47.417 15:12:37 skip_rpc.skip_rpc -- common/autotest_common.sh@959 -- # uname 00:06:47.417 15:12:37 skip_rpc.skip_rpc -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:47.417 15:12:37 skip_rpc.skip_rpc -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 69215 00:06:47.417 15:12:37 skip_rpc.skip_rpc -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:47.417 15:12:37 skip_rpc.skip_rpc -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:47.417 killing process with pid 69215 00:06:47.417 15:12:37 skip_rpc.skip_rpc -- common/autotest_common.sh@972 -- # echo 'killing process with pid 69215' 00:06:47.417 15:12:37 skip_rpc.skip_rpc -- common/autotest_common.sh@973 -- # kill 69215 00:06:47.417 15:12:37 skip_rpc.skip_rpc -- common/autotest_common.sh@978 -- # wait 69215 00:06:47.677 00:06:47.677 real 0m5.664s 00:06:47.677 user 0m5.130s 00:06:47.677 sys 0m0.462s 00:06:47.677 15:12:37 skip_rpc.skip_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:47.677 15:12:37 skip_rpc.skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:47.677 ************************************ 00:06:47.677 END TEST skip_rpc 00:06:47.677 ************************************ 00:06:47.677 15:12:37 skip_rpc -- rpc/skip_rpc.sh@74 -- # run_test skip_rpc_with_json test_skip_rpc_with_json 00:06:47.677 15:12:37 skip_rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:47.677 15:12:37 skip_rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:47.677 15:12:37 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:47.677 ************************************ 00:06:47.677 START TEST skip_rpc_with_json 00:06:47.677 ************************************ 00:06:47.677 15:12:37 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@1129 -- # test_skip_rpc_with_json 00:06:47.677 15:12:37 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@44 -- # gen_json_config 00:06:47.677 15:12:37 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@28 -- # local spdk_pid=69303 00:06:47.677 15:12:37 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@27 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:06:47.677 15:12:37 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@30 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:06:47.677 15:12:37 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@31 -- # waitforlisten 69303 00:06:47.677 15:12:37 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@835 -- # '[' -z 69303 ']' 00:06:47.677 15:12:37 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:47.677 15:12:37 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:47.677 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:47.677 15:12:37 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:47.677 15:12:37 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:47.677 15:12:37 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:06:47.677 [2024-11-19 15:12:37.978185] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:06:47.677 [2024-11-19 15:12:37.978384] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69303 ] 00:06:47.937 [2024-11-19 15:12:38.112246] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:47.937 [2024-11-19 15:12:38.155272] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:48.508 15:12:38 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:48.508 15:12:38 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@868 -- # return 0 00:06:48.508 15:12:38 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@34 -- # rpc_cmd nvmf_get_transports --trtype tcp 00:06:48.508 15:12:38 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:48.508 15:12:38 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:06:48.508 [2024-11-19 15:12:38.809076] nvmf_rpc.c:2703:rpc_nvmf_get_transports: *ERROR*: transport 'tcp' does not exist 00:06:48.508 request: 00:06:48.508 { 00:06:48.508 "trtype": "tcp", 00:06:48.508 "method": "nvmf_get_transports", 00:06:48.508 "req_id": 1 00:06:48.508 } 00:06:48.508 Got JSON-RPC error response 00:06:48.508 response: 00:06:48.508 { 00:06:48.508 "code": -19, 00:06:48.508 "message": "No such device" 00:06:48.508 } 00:06:48.508 15:12:38 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:06:48.508 15:12:38 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@34 -- # rpc_cmd nvmf_create_transport -t tcp 00:06:48.508 15:12:38 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:48.508 15:12:38 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:06:48.508 [2024-11-19 15:12:38.821215] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:06:48.508 15:12:38 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:48.508 15:12:38 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@36 -- # rpc_cmd save_config 00:06:48.508 15:12:38 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:48.508 15:12:38 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:06:48.769 15:12:38 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:48.769 15:12:38 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@37 -- # cat /home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:06:48.769 { 00:06:48.769 "subsystems": [ 00:06:48.769 { 00:06:48.769 "subsystem": "fsdev", 00:06:48.769 "config": [ 00:06:48.769 { 00:06:48.769 "method": "fsdev_set_opts", 00:06:48.769 "params": { 00:06:48.769 "fsdev_io_pool_size": 65535, 00:06:48.769 "fsdev_io_cache_size": 256 00:06:48.769 } 00:06:48.769 } 00:06:48.769 ] 00:06:48.769 }, 00:06:48.769 { 00:06:48.769 "subsystem": "keyring", 00:06:48.769 "config": [] 00:06:48.769 }, 00:06:48.769 { 00:06:48.769 "subsystem": "iobuf", 00:06:48.769 "config": [ 00:06:48.769 { 00:06:48.769 "method": "iobuf_set_options", 00:06:48.769 "params": { 00:06:48.769 "small_pool_count": 8192, 00:06:48.769 "large_pool_count": 1024, 00:06:48.769 "small_bufsize": 8192, 00:06:48.769 "large_bufsize": 135168, 00:06:48.769 "enable_numa": false 00:06:48.769 } 00:06:48.769 } 00:06:48.769 ] 00:06:48.769 }, 00:06:48.769 { 00:06:48.769 "subsystem": "sock", 00:06:48.769 "config": [ 00:06:48.769 { 00:06:48.769 "method": "sock_set_default_impl", 00:06:48.769 "params": { 00:06:48.769 "impl_name": "posix" 00:06:48.769 } 00:06:48.769 }, 00:06:48.769 { 00:06:48.769 "method": "sock_impl_set_options", 00:06:48.769 "params": { 00:06:48.769 "impl_name": "ssl", 00:06:48.769 "recv_buf_size": 4096, 00:06:48.769 "send_buf_size": 4096, 00:06:48.769 "enable_recv_pipe": true, 00:06:48.769 "enable_quickack": false, 00:06:48.769 "enable_placement_id": 0, 00:06:48.769 "enable_zerocopy_send_server": true, 00:06:48.769 "enable_zerocopy_send_client": false, 00:06:48.769 "zerocopy_threshold": 0, 00:06:48.769 "tls_version": 0, 00:06:48.769 "enable_ktls": false 00:06:48.769 } 00:06:48.769 }, 00:06:48.769 { 00:06:48.769 "method": "sock_impl_set_options", 00:06:48.769 "params": { 00:06:48.769 "impl_name": "posix", 00:06:48.769 "recv_buf_size": 2097152, 00:06:48.769 "send_buf_size": 2097152, 00:06:48.769 "enable_recv_pipe": true, 00:06:48.769 "enable_quickack": false, 00:06:48.769 "enable_placement_id": 0, 00:06:48.769 "enable_zerocopy_send_server": true, 00:06:48.769 "enable_zerocopy_send_client": false, 00:06:48.769 "zerocopy_threshold": 0, 00:06:48.769 "tls_version": 0, 00:06:48.769 "enable_ktls": false 00:06:48.769 } 00:06:48.769 } 00:06:48.769 ] 00:06:48.769 }, 00:06:48.769 { 00:06:48.769 "subsystem": "vmd", 00:06:48.769 "config": [] 00:06:48.769 }, 00:06:48.769 { 00:06:48.769 "subsystem": "accel", 00:06:48.769 "config": [ 00:06:48.769 { 00:06:48.769 "method": "accel_set_options", 00:06:48.769 "params": { 00:06:48.769 "small_cache_size": 128, 00:06:48.769 "large_cache_size": 16, 00:06:48.769 "task_count": 2048, 00:06:48.770 "sequence_count": 2048, 00:06:48.770 "buf_count": 2048 00:06:48.770 } 00:06:48.770 } 00:06:48.770 ] 00:06:48.770 }, 00:06:48.770 { 00:06:48.770 "subsystem": "bdev", 00:06:48.770 "config": [ 00:06:48.770 { 00:06:48.770 "method": "bdev_set_options", 00:06:48.770 "params": { 00:06:48.770 "bdev_io_pool_size": 65535, 00:06:48.770 "bdev_io_cache_size": 256, 00:06:48.770 "bdev_auto_examine": true, 00:06:48.770 "iobuf_small_cache_size": 128, 00:06:48.770 "iobuf_large_cache_size": 16 00:06:48.770 } 00:06:48.770 }, 00:06:48.770 { 00:06:48.770 "method": "bdev_raid_set_options", 00:06:48.770 "params": { 00:06:48.770 "process_window_size_kb": 1024, 00:06:48.770 "process_max_bandwidth_mb_sec": 0 00:06:48.770 } 00:06:48.770 }, 00:06:48.770 { 00:06:48.770 "method": "bdev_iscsi_set_options", 00:06:48.770 "params": { 00:06:48.770 "timeout_sec": 30 00:06:48.770 } 00:06:48.770 }, 00:06:48.770 { 00:06:48.770 "method": "bdev_nvme_set_options", 00:06:48.770 "params": { 00:06:48.770 "action_on_timeout": "none", 00:06:48.770 "timeout_us": 0, 00:06:48.770 "timeout_admin_us": 0, 00:06:48.770 "keep_alive_timeout_ms": 10000, 00:06:48.770 "arbitration_burst": 0, 00:06:48.770 "low_priority_weight": 0, 00:06:48.770 "medium_priority_weight": 0, 00:06:48.770 "high_priority_weight": 0, 00:06:48.770 "nvme_adminq_poll_period_us": 10000, 00:06:48.770 "nvme_ioq_poll_period_us": 0, 00:06:48.770 "io_queue_requests": 0, 00:06:48.770 "delay_cmd_submit": true, 00:06:48.770 "transport_retry_count": 4, 00:06:48.770 "bdev_retry_count": 3, 00:06:48.770 "transport_ack_timeout": 0, 00:06:48.770 "ctrlr_loss_timeout_sec": 0, 00:06:48.770 "reconnect_delay_sec": 0, 00:06:48.770 "fast_io_fail_timeout_sec": 0, 00:06:48.770 "disable_auto_failback": false, 00:06:48.770 "generate_uuids": false, 00:06:48.770 "transport_tos": 0, 00:06:48.770 "nvme_error_stat": false, 00:06:48.770 "rdma_srq_size": 0, 00:06:48.770 "io_path_stat": false, 00:06:48.770 "allow_accel_sequence": false, 00:06:48.770 "rdma_max_cq_size": 0, 00:06:48.770 "rdma_cm_event_timeout_ms": 0, 00:06:48.770 "dhchap_digests": [ 00:06:48.770 "sha256", 00:06:48.770 "sha384", 00:06:48.770 "sha512" 00:06:48.770 ], 00:06:48.770 "dhchap_dhgroups": [ 00:06:48.770 "null", 00:06:48.770 "ffdhe2048", 00:06:48.770 "ffdhe3072", 00:06:48.770 "ffdhe4096", 00:06:48.770 "ffdhe6144", 00:06:48.770 "ffdhe8192" 00:06:48.770 ] 00:06:48.770 } 00:06:48.770 }, 00:06:48.770 { 00:06:48.770 "method": "bdev_nvme_set_hotplug", 00:06:48.770 "params": { 00:06:48.770 "period_us": 100000, 00:06:48.770 "enable": false 00:06:48.770 } 00:06:48.770 }, 00:06:48.770 { 00:06:48.770 "method": "bdev_wait_for_examine" 00:06:48.770 } 00:06:48.770 ] 00:06:48.770 }, 00:06:48.770 { 00:06:48.770 "subsystem": "scsi", 00:06:48.770 "config": null 00:06:48.770 }, 00:06:48.770 { 00:06:48.770 "subsystem": "scheduler", 00:06:48.770 "config": [ 00:06:48.770 { 00:06:48.770 "method": "framework_set_scheduler", 00:06:48.770 "params": { 00:06:48.770 "name": "static" 00:06:48.770 } 00:06:48.770 } 00:06:48.770 ] 00:06:48.770 }, 00:06:48.770 { 00:06:48.770 "subsystem": "vhost_scsi", 00:06:48.770 "config": [] 00:06:48.770 }, 00:06:48.770 { 00:06:48.770 "subsystem": "vhost_blk", 00:06:48.770 "config": [] 00:06:48.770 }, 00:06:48.770 { 00:06:48.770 "subsystem": "ublk", 00:06:48.770 "config": [] 00:06:48.770 }, 00:06:48.770 { 00:06:48.770 "subsystem": "nbd", 00:06:48.770 "config": [] 00:06:48.770 }, 00:06:48.770 { 00:06:48.770 "subsystem": "nvmf", 00:06:48.770 "config": [ 00:06:48.770 { 00:06:48.770 "method": "nvmf_set_config", 00:06:48.770 "params": { 00:06:48.770 "discovery_filter": "match_any", 00:06:48.770 "admin_cmd_passthru": { 00:06:48.770 "identify_ctrlr": false 00:06:48.770 }, 00:06:48.770 "dhchap_digests": [ 00:06:48.770 "sha256", 00:06:48.770 "sha384", 00:06:48.770 "sha512" 00:06:48.770 ], 00:06:48.770 "dhchap_dhgroups": [ 00:06:48.770 "null", 00:06:48.770 "ffdhe2048", 00:06:48.770 "ffdhe3072", 00:06:48.770 "ffdhe4096", 00:06:48.770 "ffdhe6144", 00:06:48.770 "ffdhe8192" 00:06:48.770 ] 00:06:48.770 } 00:06:48.770 }, 00:06:48.770 { 00:06:48.770 "method": "nvmf_set_max_subsystems", 00:06:48.770 "params": { 00:06:48.770 "max_subsystems": 1024 00:06:48.770 } 00:06:48.770 }, 00:06:48.770 { 00:06:48.770 "method": "nvmf_set_crdt", 00:06:48.770 "params": { 00:06:48.770 "crdt1": 0, 00:06:48.770 "crdt2": 0, 00:06:48.770 "crdt3": 0 00:06:48.770 } 00:06:48.770 }, 00:06:48.770 { 00:06:48.770 "method": "nvmf_create_transport", 00:06:48.770 "params": { 00:06:48.770 "trtype": "TCP", 00:06:48.770 "max_queue_depth": 128, 00:06:48.770 "max_io_qpairs_per_ctrlr": 127, 00:06:48.770 "in_capsule_data_size": 4096, 00:06:48.770 "max_io_size": 131072, 00:06:48.770 "io_unit_size": 131072, 00:06:48.770 "max_aq_depth": 128, 00:06:48.770 "num_shared_buffers": 511, 00:06:48.770 "buf_cache_size": 4294967295, 00:06:48.770 "dif_insert_or_strip": false, 00:06:48.770 "zcopy": false, 00:06:48.770 "c2h_success": true, 00:06:48.770 "sock_priority": 0, 00:06:48.770 "abort_timeout_sec": 1, 00:06:48.770 "ack_timeout": 0, 00:06:48.770 "data_wr_pool_size": 0 00:06:48.770 } 00:06:48.770 } 00:06:48.770 ] 00:06:48.770 }, 00:06:48.770 { 00:06:48.770 "subsystem": "iscsi", 00:06:48.770 "config": [ 00:06:48.770 { 00:06:48.770 "method": "iscsi_set_options", 00:06:48.770 "params": { 00:06:48.770 "node_base": "iqn.2016-06.io.spdk", 00:06:48.770 "max_sessions": 128, 00:06:48.770 "max_connections_per_session": 2, 00:06:48.770 "max_queue_depth": 64, 00:06:48.770 "default_time2wait": 2, 00:06:48.770 "default_time2retain": 20, 00:06:48.770 "first_burst_length": 8192, 00:06:48.770 "immediate_data": true, 00:06:48.770 "allow_duplicated_isid": false, 00:06:48.770 "error_recovery_level": 0, 00:06:48.770 "nop_timeout": 60, 00:06:48.770 "nop_in_interval": 30, 00:06:48.770 "disable_chap": false, 00:06:48.770 "require_chap": false, 00:06:48.770 "mutual_chap": false, 00:06:48.770 "chap_group": 0, 00:06:48.770 "max_large_datain_per_connection": 64, 00:06:48.770 "max_r2t_per_connection": 4, 00:06:48.770 "pdu_pool_size": 36864, 00:06:48.770 "immediate_data_pool_size": 16384, 00:06:48.770 "data_out_pool_size": 2048 00:06:48.770 } 00:06:48.770 } 00:06:48.770 ] 00:06:48.770 } 00:06:48.770 ] 00:06:48.770 } 00:06:48.770 15:12:38 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@39 -- # trap - SIGINT SIGTERM EXIT 00:06:48.770 15:12:38 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@40 -- # killprocess 69303 00:06:48.770 15:12:38 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@954 -- # '[' -z 69303 ']' 00:06:48.770 15:12:38 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@958 -- # kill -0 69303 00:06:48.770 15:12:38 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@959 -- # uname 00:06:48.770 15:12:38 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:48.770 15:12:39 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 69303 00:06:48.770 killing process with pid 69303 00:06:48.770 15:12:39 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:48.770 15:12:39 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:48.770 15:12:39 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@972 -- # echo 'killing process with pid 69303' 00:06:48.770 15:12:39 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@973 -- # kill 69303 00:06:48.770 15:12:39 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@978 -- # wait 69303 00:06:49.338 15:12:39 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@47 -- # local spdk_pid=69337 00:06:49.338 15:12:39 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@46 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --json /home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:06:49.338 15:12:39 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@48 -- # sleep 5 00:06:54.615 15:12:44 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@50 -- # killprocess 69337 00:06:54.615 15:12:44 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@954 -- # '[' -z 69337 ']' 00:06:54.615 15:12:44 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@958 -- # kill -0 69337 00:06:54.615 15:12:44 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@959 -- # uname 00:06:54.615 15:12:44 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:54.615 15:12:44 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 69337 00:06:54.615 killing process with pid 69337 00:06:54.615 15:12:44 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:54.615 15:12:44 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:54.615 15:12:44 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@972 -- # echo 'killing process with pid 69337' 00:06:54.615 15:12:44 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@973 -- # kill 69337 00:06:54.615 15:12:44 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@978 -- # wait 69337 00:06:55.182 15:12:45 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@51 -- # grep -q 'TCP Transport Init' /home/vagrant/spdk_repo/spdk/test/rpc/log.txt 00:06:55.182 15:12:45 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@52 -- # rm /home/vagrant/spdk_repo/spdk/test/rpc/log.txt 00:06:55.182 ************************************ 00:06:55.182 END TEST skip_rpc_with_json 00:06:55.182 ************************************ 00:06:55.182 00:06:55.182 real 0m7.432s 00:06:55.182 user 0m6.764s 00:06:55.182 sys 0m0.968s 00:06:55.182 15:12:45 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:55.182 15:12:45 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:06:55.182 15:12:45 skip_rpc -- rpc/skip_rpc.sh@75 -- # run_test skip_rpc_with_delay test_skip_rpc_with_delay 00:06:55.182 15:12:45 skip_rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:55.182 15:12:45 skip_rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:55.182 15:12:45 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:55.182 ************************************ 00:06:55.182 START TEST skip_rpc_with_delay 00:06:55.182 ************************************ 00:06:55.182 15:12:45 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@1129 -- # test_skip_rpc_with_delay 00:06:55.182 15:12:45 skip_rpc.skip_rpc_with_delay -- rpc/skip_rpc.sh@57 -- # NOT /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:06:55.182 15:12:45 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@652 -- # local es=0 00:06:55.182 15:12:45 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@654 -- # valid_exec_arg /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:06:55.182 15:12:45 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@640 -- # local arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:55.182 15:12:45 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:55.182 15:12:45 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # type -t /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:55.182 15:12:45 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:55.182 15:12:45 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@646 -- # type -P /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:55.182 15:12:45 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:55.182 15:12:45 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@646 -- # arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:55.183 15:12:45 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@646 -- # [[ -x /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt ]] 00:06:55.183 15:12:45 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@655 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:06:55.183 [2024-11-19 15:12:45.486420] app.c: 842:spdk_app_start: *ERROR*: Cannot use '--wait-for-rpc' if no RPC server is going to be started. 00:06:55.442 15:12:45 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@655 -- # es=1 00:06:55.442 15:12:45 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:06:55.442 15:12:45 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:06:55.442 15:12:45 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:06:55.442 ************************************ 00:06:55.442 END TEST skip_rpc_with_delay 00:06:55.442 ************************************ 00:06:55.442 00:06:55.442 real 0m0.165s 00:06:55.442 user 0m0.083s 00:06:55.442 sys 0m0.081s 00:06:55.442 15:12:45 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:55.442 15:12:45 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@10 -- # set +x 00:06:55.442 15:12:45 skip_rpc -- rpc/skip_rpc.sh@77 -- # uname 00:06:55.442 15:12:45 skip_rpc -- rpc/skip_rpc.sh@77 -- # '[' Linux '!=' FreeBSD ']' 00:06:55.442 15:12:45 skip_rpc -- rpc/skip_rpc.sh@78 -- # run_test exit_on_failed_rpc_init test_exit_on_failed_rpc_init 00:06:55.442 15:12:45 skip_rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:55.442 15:12:45 skip_rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:55.442 15:12:45 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:55.442 ************************************ 00:06:55.442 START TEST exit_on_failed_rpc_init 00:06:55.442 ************************************ 00:06:55.442 15:12:45 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@1129 -- # test_exit_on_failed_rpc_init 00:06:55.442 15:12:45 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@62 -- # local spdk_pid=69448 00:06:55.442 15:12:45 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@61 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:06:55.442 15:12:45 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@63 -- # waitforlisten 69448 00:06:55.442 15:12:45 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@835 -- # '[' -z 69448 ']' 00:06:55.442 15:12:45 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:55.442 15:12:45 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:55.442 15:12:45 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:55.442 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:55.442 15:12:45 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:55.442 15:12:45 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@10 -- # set +x 00:06:55.442 [2024-11-19 15:12:45.722821] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:06:55.442 [2024-11-19 15:12:45.723045] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69448 ] 00:06:55.701 [2024-11-19 15:12:45.877568] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:55.701 [2024-11-19 15:12:45.921115] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:56.270 15:12:46 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:56.270 15:12:46 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@868 -- # return 0 00:06:56.270 15:12:46 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@65 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:06:56.270 15:12:46 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@67 -- # NOT /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x2 00:06:56.270 15:12:46 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@652 -- # local es=0 00:06:56.270 15:12:46 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@654 -- # valid_exec_arg /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x2 00:06:56.270 15:12:46 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@640 -- # local arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:56.270 15:12:46 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:56.270 15:12:46 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # type -t /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:56.270 15:12:46 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:56.270 15:12:46 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@646 -- # type -P /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:56.270 15:12:46 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:56.270 15:12:46 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@646 -- # arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:56.270 15:12:46 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@646 -- # [[ -x /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt ]] 00:06:56.270 15:12:46 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@655 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x2 00:06:56.529 [2024-11-19 15:12:46.640816] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:06:56.529 [2024-11-19 15:12:46.641026] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69461 ] 00:06:56.529 [2024-11-19 15:12:46.797361] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:56.529 [2024-11-19 15:12:46.823789] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:06:56.529 [2024-11-19 15:12:46.823988] rpc.c: 180:_spdk_rpc_listen: *ERROR*: RPC Unix domain socket path /var/tmp/spdk.sock in use. Specify another. 00:06:56.529 [2024-11-19 15:12:46.824044] rpc.c: 166:spdk_rpc_initialize: *ERROR*: Unable to start RPC service at /var/tmp/spdk.sock 00:06:56.529 [2024-11-19 15:12:46.824074] app.c:1064:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:06:56.788 15:12:46 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@655 -- # es=234 00:06:56.788 15:12:46 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:06:56.788 15:12:46 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@664 -- # es=106 00:06:56.788 15:12:46 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@665 -- # case "$es" in 00:06:56.788 15:12:46 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@672 -- # es=1 00:06:56.788 15:12:46 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:06:56.788 15:12:46 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@69 -- # trap - SIGINT SIGTERM EXIT 00:06:56.788 15:12:46 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@70 -- # killprocess 69448 00:06:56.788 15:12:46 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@954 -- # '[' -z 69448 ']' 00:06:56.788 15:12:46 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@958 -- # kill -0 69448 00:06:56.788 15:12:46 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@959 -- # uname 00:06:56.788 15:12:46 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:56.788 15:12:46 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 69448 00:06:56.788 15:12:46 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:56.788 killing process with pid 69448 00:06:56.788 15:12:46 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:56.788 15:12:46 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@972 -- # echo 'killing process with pid 69448' 00:06:56.789 15:12:46 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@973 -- # kill 69448 00:06:56.789 15:12:46 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@978 -- # wait 69448 00:06:57.355 00:06:57.355 real 0m1.932s 00:06:57.355 user 0m1.891s 00:06:57.355 sys 0m0.633s 00:06:57.355 15:12:47 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:57.355 15:12:47 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@10 -- # set +x 00:06:57.355 ************************************ 00:06:57.355 END TEST exit_on_failed_rpc_init 00:06:57.355 ************************************ 00:06:57.355 15:12:47 skip_rpc -- rpc/skip_rpc.sh@81 -- # rm /home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:06:57.355 00:06:57.355 real 0m15.708s 00:06:57.355 user 0m14.086s 00:06:57.355 sys 0m2.465s 00:06:57.355 15:12:47 skip_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:57.355 15:12:47 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:57.355 ************************************ 00:06:57.355 END TEST skip_rpc 00:06:57.355 ************************************ 00:06:57.355 15:12:47 -- spdk/autotest.sh@158 -- # run_test rpc_client /home/vagrant/spdk_repo/spdk/test/rpc_client/rpc_client.sh 00:06:57.355 15:12:47 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:57.355 15:12:47 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:57.355 15:12:47 -- common/autotest_common.sh@10 -- # set +x 00:06:57.355 ************************************ 00:06:57.355 START TEST rpc_client 00:06:57.355 ************************************ 00:06:57.355 15:12:47 rpc_client -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/rpc_client/rpc_client.sh 00:06:57.614 * Looking for test storage... 00:06:57.614 * Found test storage at /home/vagrant/spdk_repo/spdk/test/rpc_client 00:06:57.614 15:12:47 rpc_client -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:06:57.614 15:12:47 rpc_client -- common/autotest_common.sh@1693 -- # lcov --version 00:06:57.614 15:12:47 rpc_client -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:06:57.614 15:12:47 rpc_client -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:06:57.614 15:12:47 rpc_client -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:57.614 15:12:47 rpc_client -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:57.614 15:12:47 rpc_client -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:57.614 15:12:47 rpc_client -- scripts/common.sh@336 -- # IFS=.-: 00:06:57.614 15:12:47 rpc_client -- scripts/common.sh@336 -- # read -ra ver1 00:06:57.614 15:12:47 rpc_client -- scripts/common.sh@337 -- # IFS=.-: 00:06:57.614 15:12:47 rpc_client -- scripts/common.sh@337 -- # read -ra ver2 00:06:57.614 15:12:47 rpc_client -- scripts/common.sh@338 -- # local 'op=<' 00:06:57.614 15:12:47 rpc_client -- scripts/common.sh@340 -- # ver1_l=2 00:06:57.614 15:12:47 rpc_client -- scripts/common.sh@341 -- # ver2_l=1 00:06:57.614 15:12:47 rpc_client -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:57.614 15:12:47 rpc_client -- scripts/common.sh@344 -- # case "$op" in 00:06:57.614 15:12:47 rpc_client -- scripts/common.sh@345 -- # : 1 00:06:57.614 15:12:47 rpc_client -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:57.614 15:12:47 rpc_client -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:57.614 15:12:47 rpc_client -- scripts/common.sh@365 -- # decimal 1 00:06:57.614 15:12:47 rpc_client -- scripts/common.sh@353 -- # local d=1 00:06:57.614 15:12:47 rpc_client -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:57.614 15:12:47 rpc_client -- scripts/common.sh@355 -- # echo 1 00:06:57.614 15:12:47 rpc_client -- scripts/common.sh@365 -- # ver1[v]=1 00:06:57.614 15:12:47 rpc_client -- scripts/common.sh@366 -- # decimal 2 00:06:57.614 15:12:47 rpc_client -- scripts/common.sh@353 -- # local d=2 00:06:57.614 15:12:47 rpc_client -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:57.614 15:12:47 rpc_client -- scripts/common.sh@355 -- # echo 2 00:06:57.614 15:12:47 rpc_client -- scripts/common.sh@366 -- # ver2[v]=2 00:06:57.614 15:12:47 rpc_client -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:57.614 15:12:47 rpc_client -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:57.614 15:12:47 rpc_client -- scripts/common.sh@368 -- # return 0 00:06:57.614 15:12:47 rpc_client -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:57.614 15:12:47 rpc_client -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:06:57.614 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:57.614 --rc genhtml_branch_coverage=1 00:06:57.614 --rc genhtml_function_coverage=1 00:06:57.614 --rc genhtml_legend=1 00:06:57.614 --rc geninfo_all_blocks=1 00:06:57.614 --rc geninfo_unexecuted_blocks=1 00:06:57.614 00:06:57.614 ' 00:06:57.614 15:12:47 rpc_client -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:06:57.614 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:57.614 --rc genhtml_branch_coverage=1 00:06:57.614 --rc genhtml_function_coverage=1 00:06:57.614 --rc genhtml_legend=1 00:06:57.614 --rc geninfo_all_blocks=1 00:06:57.614 --rc geninfo_unexecuted_blocks=1 00:06:57.614 00:06:57.614 ' 00:06:57.614 15:12:47 rpc_client -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:06:57.614 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:57.614 --rc genhtml_branch_coverage=1 00:06:57.614 --rc genhtml_function_coverage=1 00:06:57.614 --rc genhtml_legend=1 00:06:57.614 --rc geninfo_all_blocks=1 00:06:57.614 --rc geninfo_unexecuted_blocks=1 00:06:57.614 00:06:57.614 ' 00:06:57.614 15:12:47 rpc_client -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:06:57.614 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:57.614 --rc genhtml_branch_coverage=1 00:06:57.614 --rc genhtml_function_coverage=1 00:06:57.614 --rc genhtml_legend=1 00:06:57.614 --rc geninfo_all_blocks=1 00:06:57.614 --rc geninfo_unexecuted_blocks=1 00:06:57.614 00:06:57.614 ' 00:06:57.614 15:12:47 rpc_client -- rpc_client/rpc_client.sh@10 -- # /home/vagrant/spdk_repo/spdk/test/rpc_client/rpc_client_test 00:06:57.873 OK 00:06:57.873 15:12:47 rpc_client -- rpc_client/rpc_client.sh@12 -- # trap - SIGINT SIGTERM EXIT 00:06:57.873 ************************************ 00:06:57.873 END TEST rpc_client 00:06:57.873 ************************************ 00:06:57.873 00:06:57.873 real 0m0.306s 00:06:57.873 user 0m0.162s 00:06:57.873 sys 0m0.162s 00:06:57.873 15:12:47 rpc_client -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:57.873 15:12:47 rpc_client -- common/autotest_common.sh@10 -- # set +x 00:06:57.873 15:12:48 -- spdk/autotest.sh@159 -- # run_test json_config /home/vagrant/spdk_repo/spdk/test/json_config/json_config.sh 00:06:57.873 15:12:48 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:57.873 15:12:48 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:57.873 15:12:48 -- common/autotest_common.sh@10 -- # set +x 00:06:57.873 ************************************ 00:06:57.873 START TEST json_config 00:06:57.873 ************************************ 00:06:57.873 15:12:48 json_config -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/json_config/json_config.sh 00:06:57.873 15:12:48 json_config -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:06:57.873 15:12:48 json_config -- common/autotest_common.sh@1693 -- # lcov --version 00:06:57.873 15:12:48 json_config -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:06:58.132 15:12:48 json_config -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:06:58.132 15:12:48 json_config -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:58.132 15:12:48 json_config -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:58.132 15:12:48 json_config -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:58.132 15:12:48 json_config -- scripts/common.sh@336 -- # IFS=.-: 00:06:58.132 15:12:48 json_config -- scripts/common.sh@336 -- # read -ra ver1 00:06:58.132 15:12:48 json_config -- scripts/common.sh@337 -- # IFS=.-: 00:06:58.132 15:12:48 json_config -- scripts/common.sh@337 -- # read -ra ver2 00:06:58.132 15:12:48 json_config -- scripts/common.sh@338 -- # local 'op=<' 00:06:58.132 15:12:48 json_config -- scripts/common.sh@340 -- # ver1_l=2 00:06:58.132 15:12:48 json_config -- scripts/common.sh@341 -- # ver2_l=1 00:06:58.132 15:12:48 json_config -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:58.132 15:12:48 json_config -- scripts/common.sh@344 -- # case "$op" in 00:06:58.132 15:12:48 json_config -- scripts/common.sh@345 -- # : 1 00:06:58.132 15:12:48 json_config -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:58.132 15:12:48 json_config -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:58.132 15:12:48 json_config -- scripts/common.sh@365 -- # decimal 1 00:06:58.132 15:12:48 json_config -- scripts/common.sh@353 -- # local d=1 00:06:58.132 15:12:48 json_config -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:58.132 15:12:48 json_config -- scripts/common.sh@355 -- # echo 1 00:06:58.132 15:12:48 json_config -- scripts/common.sh@365 -- # ver1[v]=1 00:06:58.132 15:12:48 json_config -- scripts/common.sh@366 -- # decimal 2 00:06:58.132 15:12:48 json_config -- scripts/common.sh@353 -- # local d=2 00:06:58.132 15:12:48 json_config -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:58.132 15:12:48 json_config -- scripts/common.sh@355 -- # echo 2 00:06:58.132 15:12:48 json_config -- scripts/common.sh@366 -- # ver2[v]=2 00:06:58.132 15:12:48 json_config -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:58.132 15:12:48 json_config -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:58.132 15:12:48 json_config -- scripts/common.sh@368 -- # return 0 00:06:58.132 15:12:48 json_config -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:58.132 15:12:48 json_config -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:06:58.132 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:58.132 --rc genhtml_branch_coverage=1 00:06:58.132 --rc genhtml_function_coverage=1 00:06:58.132 --rc genhtml_legend=1 00:06:58.132 --rc geninfo_all_blocks=1 00:06:58.132 --rc geninfo_unexecuted_blocks=1 00:06:58.132 00:06:58.132 ' 00:06:58.132 15:12:48 json_config -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:06:58.132 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:58.132 --rc genhtml_branch_coverage=1 00:06:58.132 --rc genhtml_function_coverage=1 00:06:58.132 --rc genhtml_legend=1 00:06:58.132 --rc geninfo_all_blocks=1 00:06:58.132 --rc geninfo_unexecuted_blocks=1 00:06:58.132 00:06:58.132 ' 00:06:58.132 15:12:48 json_config -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:06:58.132 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:58.132 --rc genhtml_branch_coverage=1 00:06:58.132 --rc genhtml_function_coverage=1 00:06:58.132 --rc genhtml_legend=1 00:06:58.132 --rc geninfo_all_blocks=1 00:06:58.132 --rc geninfo_unexecuted_blocks=1 00:06:58.132 00:06:58.133 ' 00:06:58.133 15:12:48 json_config -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:06:58.133 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:58.133 --rc genhtml_branch_coverage=1 00:06:58.133 --rc genhtml_function_coverage=1 00:06:58.133 --rc genhtml_legend=1 00:06:58.133 --rc geninfo_all_blocks=1 00:06:58.133 --rc geninfo_unexecuted_blocks=1 00:06:58.133 00:06:58.133 ' 00:06:58.133 15:12:48 json_config -- json_config/json_config.sh@8 -- # source /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh 00:06:58.133 15:12:48 json_config -- nvmf/common.sh@7 -- # uname -s 00:06:58.133 15:12:48 json_config -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:06:58.133 15:12:48 json_config -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:06:58.133 15:12:48 json_config -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:06:58.133 15:12:48 json_config -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:06:58.133 15:12:48 json_config -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:06:58.133 15:12:48 json_config -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:06:58.133 15:12:48 json_config -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:06:58.133 15:12:48 json_config -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:06:58.133 15:12:48 json_config -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:06:58.133 15:12:48 json_config -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:06:58.133 15:12:48 json_config -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:05233599-7658-47bb-b32b-d880580825e2 00:06:58.133 15:12:48 json_config -- nvmf/common.sh@18 -- # NVME_HOSTID=05233599-7658-47bb-b32b-d880580825e2 00:06:58.133 15:12:48 json_config -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:06:58.133 15:12:48 json_config -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:06:58.133 15:12:48 json_config -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:06:58.133 15:12:48 json_config -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:06:58.133 15:12:48 json_config -- nvmf/common.sh@49 -- # source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:06:58.133 15:12:48 json_config -- scripts/common.sh@15 -- # shopt -s extglob 00:06:58.133 15:12:48 json_config -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:06:58.133 15:12:48 json_config -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:06:58.133 15:12:48 json_config -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:06:58.133 15:12:48 json_config -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:58.133 15:12:48 json_config -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:58.133 15:12:48 json_config -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:58.133 15:12:48 json_config -- paths/export.sh@5 -- # export PATH 00:06:58.133 15:12:48 json_config -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:58.133 15:12:48 json_config -- nvmf/common.sh@51 -- # : 0 00:06:58.133 15:12:48 json_config -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:06:58.133 15:12:48 json_config -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:06:58.133 15:12:48 json_config -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:06:58.133 15:12:48 json_config -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:06:58.133 15:12:48 json_config -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:06:58.133 15:12:48 json_config -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:06:58.133 /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:06:58.133 15:12:48 json_config -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:06:58.133 15:12:48 json_config -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:06:58.133 15:12:48 json_config -- nvmf/common.sh@55 -- # have_pci_nics=0 00:06:58.133 15:12:48 json_config -- json_config/json_config.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/json_config/common.sh 00:06:58.133 15:12:48 json_config -- json_config/json_config.sh@11 -- # [[ 0 -eq 1 ]] 00:06:58.133 WARNING: No tests are enabled so not running JSON configuration tests 00:06:58.133 15:12:48 json_config -- json_config/json_config.sh@15 -- # [[ 0 -ne 1 ]] 00:06:58.133 15:12:48 json_config -- json_config/json_config.sh@15 -- # [[ 0 -eq 1 ]] 00:06:58.133 15:12:48 json_config -- json_config/json_config.sh@26 -- # (( SPDK_TEST_BLOCKDEV + SPDK_TEST_ISCSI + SPDK_TEST_NVMF + SPDK_TEST_VHOST + SPDK_TEST_VHOST_INIT + SPDK_TEST_RBD == 0 )) 00:06:58.133 15:12:48 json_config -- json_config/json_config.sh@27 -- # echo 'WARNING: No tests are enabled so not running JSON configuration tests' 00:06:58.133 15:12:48 json_config -- json_config/json_config.sh@28 -- # exit 0 00:06:58.133 ************************************ 00:06:58.133 END TEST json_config 00:06:58.133 ************************************ 00:06:58.133 00:06:58.133 real 0m0.232s 00:06:58.133 user 0m0.143s 00:06:58.133 sys 0m0.094s 00:06:58.133 15:12:48 json_config -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:58.133 15:12:48 json_config -- common/autotest_common.sh@10 -- # set +x 00:06:58.133 15:12:48 -- spdk/autotest.sh@160 -- # run_test json_config_extra_key /home/vagrant/spdk_repo/spdk/test/json_config/json_config_extra_key.sh 00:06:58.133 15:12:48 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:58.133 15:12:48 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:58.133 15:12:48 -- common/autotest_common.sh@10 -- # set +x 00:06:58.133 ************************************ 00:06:58.133 START TEST json_config_extra_key 00:06:58.133 ************************************ 00:06:58.133 15:12:48 json_config_extra_key -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/json_config/json_config_extra_key.sh 00:06:58.133 15:12:48 json_config_extra_key -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:06:58.133 15:12:48 json_config_extra_key -- common/autotest_common.sh@1693 -- # lcov --version 00:06:58.133 15:12:48 json_config_extra_key -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:06:58.393 15:12:48 json_config_extra_key -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:06:58.393 15:12:48 json_config_extra_key -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:58.393 15:12:48 json_config_extra_key -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:58.393 15:12:48 json_config_extra_key -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:58.393 15:12:48 json_config_extra_key -- scripts/common.sh@336 -- # IFS=.-: 00:06:58.393 15:12:48 json_config_extra_key -- scripts/common.sh@336 -- # read -ra ver1 00:06:58.393 15:12:48 json_config_extra_key -- scripts/common.sh@337 -- # IFS=.-: 00:06:58.393 15:12:48 json_config_extra_key -- scripts/common.sh@337 -- # read -ra ver2 00:06:58.393 15:12:48 json_config_extra_key -- scripts/common.sh@338 -- # local 'op=<' 00:06:58.393 15:12:48 json_config_extra_key -- scripts/common.sh@340 -- # ver1_l=2 00:06:58.393 15:12:48 json_config_extra_key -- scripts/common.sh@341 -- # ver2_l=1 00:06:58.393 15:12:48 json_config_extra_key -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:58.393 15:12:48 json_config_extra_key -- scripts/common.sh@344 -- # case "$op" in 00:06:58.393 15:12:48 json_config_extra_key -- scripts/common.sh@345 -- # : 1 00:06:58.393 15:12:48 json_config_extra_key -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:58.393 15:12:48 json_config_extra_key -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:58.393 15:12:48 json_config_extra_key -- scripts/common.sh@365 -- # decimal 1 00:06:58.393 15:12:48 json_config_extra_key -- scripts/common.sh@353 -- # local d=1 00:06:58.393 15:12:48 json_config_extra_key -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:58.393 15:12:48 json_config_extra_key -- scripts/common.sh@355 -- # echo 1 00:06:58.393 15:12:48 json_config_extra_key -- scripts/common.sh@365 -- # ver1[v]=1 00:06:58.393 15:12:48 json_config_extra_key -- scripts/common.sh@366 -- # decimal 2 00:06:58.393 15:12:48 json_config_extra_key -- scripts/common.sh@353 -- # local d=2 00:06:58.393 15:12:48 json_config_extra_key -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:58.393 15:12:48 json_config_extra_key -- scripts/common.sh@355 -- # echo 2 00:06:58.393 15:12:48 json_config_extra_key -- scripts/common.sh@366 -- # ver2[v]=2 00:06:58.393 15:12:48 json_config_extra_key -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:58.393 15:12:48 json_config_extra_key -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:58.393 15:12:48 json_config_extra_key -- scripts/common.sh@368 -- # return 0 00:06:58.393 15:12:48 json_config_extra_key -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:58.393 15:12:48 json_config_extra_key -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:06:58.393 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:58.393 --rc genhtml_branch_coverage=1 00:06:58.393 --rc genhtml_function_coverage=1 00:06:58.393 --rc genhtml_legend=1 00:06:58.393 --rc geninfo_all_blocks=1 00:06:58.393 --rc geninfo_unexecuted_blocks=1 00:06:58.393 00:06:58.393 ' 00:06:58.393 15:12:48 json_config_extra_key -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:06:58.393 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:58.393 --rc genhtml_branch_coverage=1 00:06:58.393 --rc genhtml_function_coverage=1 00:06:58.393 --rc genhtml_legend=1 00:06:58.393 --rc geninfo_all_blocks=1 00:06:58.393 --rc geninfo_unexecuted_blocks=1 00:06:58.393 00:06:58.393 ' 00:06:58.393 15:12:48 json_config_extra_key -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:06:58.393 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:58.393 --rc genhtml_branch_coverage=1 00:06:58.393 --rc genhtml_function_coverage=1 00:06:58.393 --rc genhtml_legend=1 00:06:58.393 --rc geninfo_all_blocks=1 00:06:58.393 --rc geninfo_unexecuted_blocks=1 00:06:58.393 00:06:58.393 ' 00:06:58.393 15:12:48 json_config_extra_key -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:06:58.393 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:58.393 --rc genhtml_branch_coverage=1 00:06:58.393 --rc genhtml_function_coverage=1 00:06:58.393 --rc genhtml_legend=1 00:06:58.393 --rc geninfo_all_blocks=1 00:06:58.393 --rc geninfo_unexecuted_blocks=1 00:06:58.393 00:06:58.393 ' 00:06:58.393 15:12:48 json_config_extra_key -- json_config/json_config_extra_key.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh 00:06:58.393 15:12:48 json_config_extra_key -- nvmf/common.sh@7 -- # uname -s 00:06:58.393 15:12:48 json_config_extra_key -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:06:58.393 15:12:48 json_config_extra_key -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:06:58.394 15:12:48 json_config_extra_key -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:06:58.394 15:12:48 json_config_extra_key -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:06:58.394 15:12:48 json_config_extra_key -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:06:58.394 15:12:48 json_config_extra_key -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:06:58.394 15:12:48 json_config_extra_key -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:06:58.394 15:12:48 json_config_extra_key -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:06:58.394 15:12:48 json_config_extra_key -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:06:58.394 15:12:48 json_config_extra_key -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:06:58.394 15:12:48 json_config_extra_key -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:05233599-7658-47bb-b32b-d880580825e2 00:06:58.394 15:12:48 json_config_extra_key -- nvmf/common.sh@18 -- # NVME_HOSTID=05233599-7658-47bb-b32b-d880580825e2 00:06:58.394 15:12:48 json_config_extra_key -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:06:58.394 15:12:48 json_config_extra_key -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:06:58.394 15:12:48 json_config_extra_key -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:06:58.394 15:12:48 json_config_extra_key -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:06:58.394 15:12:48 json_config_extra_key -- nvmf/common.sh@49 -- # source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:06:58.394 15:12:48 json_config_extra_key -- scripts/common.sh@15 -- # shopt -s extglob 00:06:58.394 15:12:48 json_config_extra_key -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:06:58.394 15:12:48 json_config_extra_key -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:06:58.394 15:12:48 json_config_extra_key -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:06:58.394 15:12:48 json_config_extra_key -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:58.394 15:12:48 json_config_extra_key -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:58.394 15:12:48 json_config_extra_key -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:58.394 15:12:48 json_config_extra_key -- paths/export.sh@5 -- # export PATH 00:06:58.394 15:12:48 json_config_extra_key -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:58.394 15:12:48 json_config_extra_key -- nvmf/common.sh@51 -- # : 0 00:06:58.394 15:12:48 json_config_extra_key -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:06:58.394 15:12:48 json_config_extra_key -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:06:58.394 15:12:48 json_config_extra_key -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:06:58.394 15:12:48 json_config_extra_key -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:06:58.394 15:12:48 json_config_extra_key -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:06:58.394 15:12:48 json_config_extra_key -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:06:58.394 /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:06:58.394 15:12:48 json_config_extra_key -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:06:58.394 15:12:48 json_config_extra_key -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:06:58.394 15:12:48 json_config_extra_key -- nvmf/common.sh@55 -- # have_pci_nics=0 00:06:58.394 15:12:48 json_config_extra_key -- json_config/json_config_extra_key.sh@10 -- # source /home/vagrant/spdk_repo/spdk/test/json_config/common.sh 00:06:58.394 15:12:48 json_config_extra_key -- json_config/json_config_extra_key.sh@17 -- # app_pid=(['target']='') 00:06:58.394 15:12:48 json_config_extra_key -- json_config/json_config_extra_key.sh@17 -- # declare -A app_pid 00:06:58.394 15:12:48 json_config_extra_key -- json_config/json_config_extra_key.sh@18 -- # app_socket=(['target']='/var/tmp/spdk_tgt.sock') 00:06:58.394 15:12:48 json_config_extra_key -- json_config/json_config_extra_key.sh@18 -- # declare -A app_socket 00:06:58.394 15:12:48 json_config_extra_key -- json_config/json_config_extra_key.sh@19 -- # app_params=(['target']='-m 0x1 -s 1024') 00:06:58.394 15:12:48 json_config_extra_key -- json_config/json_config_extra_key.sh@19 -- # declare -A app_params 00:06:58.394 15:12:48 json_config_extra_key -- json_config/json_config_extra_key.sh@20 -- # configs_path=(['target']='/home/vagrant/spdk_repo/spdk/test/json_config/extra_key.json') 00:06:58.394 15:12:48 json_config_extra_key -- json_config/json_config_extra_key.sh@20 -- # declare -A configs_path 00:06:58.394 15:12:48 json_config_extra_key -- json_config/json_config_extra_key.sh@22 -- # trap 'on_error_exit "${FUNCNAME}" "${LINENO}"' ERR 00:06:58.394 15:12:48 json_config_extra_key -- json_config/json_config_extra_key.sh@24 -- # echo 'INFO: launching applications...' 00:06:58.394 INFO: launching applications... 00:06:58.394 15:12:48 json_config_extra_key -- json_config/json_config_extra_key.sh@25 -- # json_config_test_start_app target --json /home/vagrant/spdk_repo/spdk/test/json_config/extra_key.json 00:06:58.394 15:12:48 json_config_extra_key -- json_config/common.sh@9 -- # local app=target 00:06:58.394 15:12:48 json_config_extra_key -- json_config/common.sh@10 -- # shift 00:06:58.394 15:12:48 json_config_extra_key -- json_config/common.sh@12 -- # [[ -n 22 ]] 00:06:58.394 15:12:48 json_config_extra_key -- json_config/common.sh@13 -- # [[ -z '' ]] 00:06:58.394 15:12:48 json_config_extra_key -- json_config/common.sh@15 -- # local app_extra_params= 00:06:58.394 15:12:48 json_config_extra_key -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:06:58.394 15:12:48 json_config_extra_key -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:06:58.394 15:12:48 json_config_extra_key -- json_config/common.sh@22 -- # app_pid["$app"]=69649 00:06:58.394 15:12:48 json_config_extra_key -- json_config/common.sh@21 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 -s 1024 -r /var/tmp/spdk_tgt.sock --json /home/vagrant/spdk_repo/spdk/test/json_config/extra_key.json 00:06:58.394 15:12:48 json_config_extra_key -- json_config/common.sh@24 -- # echo 'Waiting for target to run...' 00:06:58.394 Waiting for target to run... 00:06:58.394 15:12:48 json_config_extra_key -- json_config/common.sh@25 -- # waitforlisten 69649 /var/tmp/spdk_tgt.sock 00:06:58.394 15:12:48 json_config_extra_key -- common/autotest_common.sh@835 -- # '[' -z 69649 ']' 00:06:58.394 15:12:48 json_config_extra_key -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk_tgt.sock 00:06:58.394 15:12:48 json_config_extra_key -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:58.394 15:12:48 json_config_extra_key -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock...' 00:06:58.394 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock... 00:06:58.394 15:12:48 json_config_extra_key -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:58.394 15:12:48 json_config_extra_key -- common/autotest_common.sh@10 -- # set +x 00:06:58.394 [2024-11-19 15:12:48.676712] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:06:58.394 [2024-11-19 15:12:48.676959] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 -m 1024 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69649 ] 00:06:58.964 [2024-11-19 15:12:49.051023] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:58.964 [2024-11-19 15:12:49.074911] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:59.223 15:12:49 json_config_extra_key -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:59.223 15:12:49 json_config_extra_key -- common/autotest_common.sh@868 -- # return 0 00:06:59.223 15:12:49 json_config_extra_key -- json_config/common.sh@26 -- # echo '' 00:06:59.223 00:06:59.223 15:12:49 json_config_extra_key -- json_config/json_config_extra_key.sh@27 -- # echo 'INFO: shutting down applications...' 00:06:59.223 INFO: shutting down applications... 00:06:59.224 15:12:49 json_config_extra_key -- json_config/json_config_extra_key.sh@28 -- # json_config_test_shutdown_app target 00:06:59.224 15:12:49 json_config_extra_key -- json_config/common.sh@31 -- # local app=target 00:06:59.224 15:12:49 json_config_extra_key -- json_config/common.sh@34 -- # [[ -n 22 ]] 00:06:59.224 15:12:49 json_config_extra_key -- json_config/common.sh@35 -- # [[ -n 69649 ]] 00:06:59.224 15:12:49 json_config_extra_key -- json_config/common.sh@38 -- # kill -SIGINT 69649 00:06:59.224 15:12:49 json_config_extra_key -- json_config/common.sh@40 -- # (( i = 0 )) 00:06:59.224 15:12:49 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:06:59.224 15:12:49 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 69649 00:06:59.224 15:12:49 json_config_extra_key -- json_config/common.sh@45 -- # sleep 0.5 00:06:59.792 15:12:50 json_config_extra_key -- json_config/common.sh@40 -- # (( i++ )) 00:06:59.792 15:12:50 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:06:59.792 15:12:50 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 69649 00:06:59.792 15:12:50 json_config_extra_key -- json_config/common.sh@45 -- # sleep 0.5 00:07:00.361 15:12:50 json_config_extra_key -- json_config/common.sh@40 -- # (( i++ )) 00:07:00.361 15:12:50 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:07:00.361 15:12:50 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 69649 00:07:00.361 SPDK target shutdown done 00:07:00.361 Success 00:07:00.361 15:12:50 json_config_extra_key -- json_config/common.sh@42 -- # app_pid["$app"]= 00:07:00.361 15:12:50 json_config_extra_key -- json_config/common.sh@43 -- # break 00:07:00.361 15:12:50 json_config_extra_key -- json_config/common.sh@48 -- # [[ -n '' ]] 00:07:00.361 15:12:50 json_config_extra_key -- json_config/common.sh@53 -- # echo 'SPDK target shutdown done' 00:07:00.361 15:12:50 json_config_extra_key -- json_config/json_config_extra_key.sh@30 -- # echo Success 00:07:00.361 00:07:00.361 real 0m2.203s 00:07:00.361 user 0m1.692s 00:07:00.361 sys 0m0.494s 00:07:00.361 15:12:50 json_config_extra_key -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:00.361 15:12:50 json_config_extra_key -- common/autotest_common.sh@10 -- # set +x 00:07:00.361 ************************************ 00:07:00.361 END TEST json_config_extra_key 00:07:00.361 ************************************ 00:07:00.361 15:12:50 -- spdk/autotest.sh@161 -- # run_test alias_rpc /home/vagrant/spdk_repo/spdk/test/json_config/alias_rpc/alias_rpc.sh 00:07:00.361 15:12:50 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:07:00.361 15:12:50 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:00.361 15:12:50 -- common/autotest_common.sh@10 -- # set +x 00:07:00.361 ************************************ 00:07:00.361 START TEST alias_rpc 00:07:00.361 ************************************ 00:07:00.361 15:12:50 alias_rpc -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/json_config/alias_rpc/alias_rpc.sh 00:07:00.620 * Looking for test storage... 00:07:00.620 * Found test storage at /home/vagrant/spdk_repo/spdk/test/json_config/alias_rpc 00:07:00.620 15:12:50 alias_rpc -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:07:00.620 15:12:50 alias_rpc -- common/autotest_common.sh@1693 -- # lcov --version 00:07:00.620 15:12:50 alias_rpc -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:07:00.620 15:12:50 alias_rpc -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:07:00.620 15:12:50 alias_rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:07:00.620 15:12:50 alias_rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:07:00.620 15:12:50 alias_rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:07:00.620 15:12:50 alias_rpc -- scripts/common.sh@336 -- # IFS=.-: 00:07:00.621 15:12:50 alias_rpc -- scripts/common.sh@336 -- # read -ra ver1 00:07:00.621 15:12:50 alias_rpc -- scripts/common.sh@337 -- # IFS=.-: 00:07:00.621 15:12:50 alias_rpc -- scripts/common.sh@337 -- # read -ra ver2 00:07:00.621 15:12:50 alias_rpc -- scripts/common.sh@338 -- # local 'op=<' 00:07:00.621 15:12:50 alias_rpc -- scripts/common.sh@340 -- # ver1_l=2 00:07:00.621 15:12:50 alias_rpc -- scripts/common.sh@341 -- # ver2_l=1 00:07:00.621 15:12:50 alias_rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:07:00.621 15:12:50 alias_rpc -- scripts/common.sh@344 -- # case "$op" in 00:07:00.621 15:12:50 alias_rpc -- scripts/common.sh@345 -- # : 1 00:07:00.621 15:12:50 alias_rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:07:00.621 15:12:50 alias_rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:07:00.621 15:12:50 alias_rpc -- scripts/common.sh@365 -- # decimal 1 00:07:00.621 15:12:50 alias_rpc -- scripts/common.sh@353 -- # local d=1 00:07:00.621 15:12:50 alias_rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:07:00.621 15:12:50 alias_rpc -- scripts/common.sh@355 -- # echo 1 00:07:00.621 15:12:50 alias_rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:07:00.621 15:12:50 alias_rpc -- scripts/common.sh@366 -- # decimal 2 00:07:00.621 15:12:50 alias_rpc -- scripts/common.sh@353 -- # local d=2 00:07:00.621 15:12:50 alias_rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:07:00.621 15:12:50 alias_rpc -- scripts/common.sh@355 -- # echo 2 00:07:00.621 15:12:50 alias_rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:07:00.621 15:12:50 alias_rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:07:00.621 15:12:50 alias_rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:07:00.621 15:12:50 alias_rpc -- scripts/common.sh@368 -- # return 0 00:07:00.621 15:12:50 alias_rpc -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:07:00.621 15:12:50 alias_rpc -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:07:00.621 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:00.621 --rc genhtml_branch_coverage=1 00:07:00.621 --rc genhtml_function_coverage=1 00:07:00.621 --rc genhtml_legend=1 00:07:00.621 --rc geninfo_all_blocks=1 00:07:00.621 --rc geninfo_unexecuted_blocks=1 00:07:00.621 00:07:00.621 ' 00:07:00.621 15:12:50 alias_rpc -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:07:00.621 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:00.621 --rc genhtml_branch_coverage=1 00:07:00.621 --rc genhtml_function_coverage=1 00:07:00.621 --rc genhtml_legend=1 00:07:00.621 --rc geninfo_all_blocks=1 00:07:00.621 --rc geninfo_unexecuted_blocks=1 00:07:00.621 00:07:00.621 ' 00:07:00.621 15:12:50 alias_rpc -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:07:00.621 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:00.621 --rc genhtml_branch_coverage=1 00:07:00.621 --rc genhtml_function_coverage=1 00:07:00.621 --rc genhtml_legend=1 00:07:00.621 --rc geninfo_all_blocks=1 00:07:00.621 --rc geninfo_unexecuted_blocks=1 00:07:00.621 00:07:00.621 ' 00:07:00.621 15:12:50 alias_rpc -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:07:00.621 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:00.621 --rc genhtml_branch_coverage=1 00:07:00.621 --rc genhtml_function_coverage=1 00:07:00.621 --rc genhtml_legend=1 00:07:00.621 --rc geninfo_all_blocks=1 00:07:00.621 --rc geninfo_unexecuted_blocks=1 00:07:00.621 00:07:00.621 ' 00:07:00.621 15:12:50 alias_rpc -- alias_rpc/alias_rpc.sh@10 -- # trap 'killprocess $spdk_tgt_pid; exit 1' ERR 00:07:00.621 15:12:50 alias_rpc -- alias_rpc/alias_rpc.sh@13 -- # spdk_tgt_pid=69735 00:07:00.621 15:12:50 alias_rpc -- alias_rpc/alias_rpc.sh@12 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:07:00.621 15:12:50 alias_rpc -- alias_rpc/alias_rpc.sh@14 -- # waitforlisten 69735 00:07:00.621 15:12:50 alias_rpc -- common/autotest_common.sh@835 -- # '[' -z 69735 ']' 00:07:00.621 15:12:50 alias_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:00.621 15:12:50 alias_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:00.621 15:12:50 alias_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:00.621 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:00.621 15:12:50 alias_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:00.621 15:12:50 alias_rpc -- common/autotest_common.sh@10 -- # set +x 00:07:00.880 [2024-11-19 15:12:50.978324] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:07:00.880 [2024-11-19 15:12:50.978636] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69735 ] 00:07:00.880 [2024-11-19 15:12:51.135195] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:00.880 [2024-11-19 15:12:51.179013] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:01.820 15:12:51 alias_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:01.820 15:12:51 alias_rpc -- common/autotest_common.sh@868 -- # return 0 00:07:01.820 15:12:51 alias_rpc -- alias_rpc/alias_rpc.sh@17 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py load_config -i 00:07:01.820 15:12:52 alias_rpc -- alias_rpc/alias_rpc.sh@19 -- # killprocess 69735 00:07:01.820 15:12:52 alias_rpc -- common/autotest_common.sh@954 -- # '[' -z 69735 ']' 00:07:01.820 15:12:52 alias_rpc -- common/autotest_common.sh@958 -- # kill -0 69735 00:07:01.820 15:12:52 alias_rpc -- common/autotest_common.sh@959 -- # uname 00:07:01.820 15:12:52 alias_rpc -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:01.820 15:12:52 alias_rpc -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 69735 00:07:01.820 killing process with pid 69735 00:07:01.820 15:12:52 alias_rpc -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:01.820 15:12:52 alias_rpc -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:01.820 15:12:52 alias_rpc -- common/autotest_common.sh@972 -- # echo 'killing process with pid 69735' 00:07:01.820 15:12:52 alias_rpc -- common/autotest_common.sh@973 -- # kill 69735 00:07:01.820 15:12:52 alias_rpc -- common/autotest_common.sh@978 -- # wait 69735 00:07:02.401 ************************************ 00:07:02.401 END TEST alias_rpc 00:07:02.401 ************************************ 00:07:02.401 00:07:02.401 real 0m2.055s 00:07:02.401 user 0m1.963s 00:07:02.401 sys 0m0.665s 00:07:02.401 15:12:52 alias_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:02.401 15:12:52 alias_rpc -- common/autotest_common.sh@10 -- # set +x 00:07:02.401 15:12:52 -- spdk/autotest.sh@163 -- # [[ 0 -eq 0 ]] 00:07:02.401 15:12:52 -- spdk/autotest.sh@164 -- # run_test spdkcli_tcp /home/vagrant/spdk_repo/spdk/test/spdkcli/tcp.sh 00:07:02.401 15:12:52 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:07:02.401 15:12:52 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:02.401 15:12:52 -- common/autotest_common.sh@10 -- # set +x 00:07:02.661 ************************************ 00:07:02.661 START TEST spdkcli_tcp 00:07:02.661 ************************************ 00:07:02.661 15:12:52 spdkcli_tcp -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/tcp.sh 00:07:02.661 * Looking for test storage... 00:07:02.661 * Found test storage at /home/vagrant/spdk_repo/spdk/test/spdkcli 00:07:02.661 15:12:52 spdkcli_tcp -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:07:02.661 15:12:52 spdkcli_tcp -- common/autotest_common.sh@1693 -- # lcov --version 00:07:02.661 15:12:52 spdkcli_tcp -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:07:02.661 15:12:52 spdkcli_tcp -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:07:02.661 15:12:52 spdkcli_tcp -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:07:02.661 15:12:52 spdkcli_tcp -- scripts/common.sh@333 -- # local ver1 ver1_l 00:07:02.661 15:12:52 spdkcli_tcp -- scripts/common.sh@334 -- # local ver2 ver2_l 00:07:02.661 15:12:52 spdkcli_tcp -- scripts/common.sh@336 -- # IFS=.-: 00:07:02.661 15:12:52 spdkcli_tcp -- scripts/common.sh@336 -- # read -ra ver1 00:07:02.661 15:12:52 spdkcli_tcp -- scripts/common.sh@337 -- # IFS=.-: 00:07:02.661 15:12:52 spdkcli_tcp -- scripts/common.sh@337 -- # read -ra ver2 00:07:02.661 15:12:52 spdkcli_tcp -- scripts/common.sh@338 -- # local 'op=<' 00:07:02.661 15:12:52 spdkcli_tcp -- scripts/common.sh@340 -- # ver1_l=2 00:07:02.661 15:12:52 spdkcli_tcp -- scripts/common.sh@341 -- # ver2_l=1 00:07:02.661 15:12:52 spdkcli_tcp -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:07:02.661 15:12:52 spdkcli_tcp -- scripts/common.sh@344 -- # case "$op" in 00:07:02.661 15:12:52 spdkcli_tcp -- scripts/common.sh@345 -- # : 1 00:07:02.661 15:12:52 spdkcli_tcp -- scripts/common.sh@364 -- # (( v = 0 )) 00:07:02.661 15:12:52 spdkcli_tcp -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:07:02.661 15:12:52 spdkcli_tcp -- scripts/common.sh@365 -- # decimal 1 00:07:02.661 15:12:52 spdkcli_tcp -- scripts/common.sh@353 -- # local d=1 00:07:02.661 15:12:52 spdkcli_tcp -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:07:02.661 15:12:52 spdkcli_tcp -- scripts/common.sh@355 -- # echo 1 00:07:02.661 15:12:52 spdkcli_tcp -- scripts/common.sh@365 -- # ver1[v]=1 00:07:02.661 15:12:52 spdkcli_tcp -- scripts/common.sh@366 -- # decimal 2 00:07:02.661 15:12:52 spdkcli_tcp -- scripts/common.sh@353 -- # local d=2 00:07:02.661 15:12:52 spdkcli_tcp -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:07:02.661 15:12:52 spdkcli_tcp -- scripts/common.sh@355 -- # echo 2 00:07:02.661 15:12:52 spdkcli_tcp -- scripts/common.sh@366 -- # ver2[v]=2 00:07:02.661 15:12:52 spdkcli_tcp -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:07:02.661 15:12:52 spdkcli_tcp -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:07:02.661 15:12:52 spdkcli_tcp -- scripts/common.sh@368 -- # return 0 00:07:02.661 15:12:52 spdkcli_tcp -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:07:02.661 15:12:52 spdkcli_tcp -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:07:02.661 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:02.661 --rc genhtml_branch_coverage=1 00:07:02.661 --rc genhtml_function_coverage=1 00:07:02.661 --rc genhtml_legend=1 00:07:02.661 --rc geninfo_all_blocks=1 00:07:02.661 --rc geninfo_unexecuted_blocks=1 00:07:02.661 00:07:02.661 ' 00:07:02.661 15:12:52 spdkcli_tcp -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:07:02.661 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:02.661 --rc genhtml_branch_coverage=1 00:07:02.661 --rc genhtml_function_coverage=1 00:07:02.661 --rc genhtml_legend=1 00:07:02.661 --rc geninfo_all_blocks=1 00:07:02.661 --rc geninfo_unexecuted_blocks=1 00:07:02.661 00:07:02.661 ' 00:07:02.661 15:12:52 spdkcli_tcp -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:07:02.661 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:02.661 --rc genhtml_branch_coverage=1 00:07:02.661 --rc genhtml_function_coverage=1 00:07:02.661 --rc genhtml_legend=1 00:07:02.661 --rc geninfo_all_blocks=1 00:07:02.661 --rc geninfo_unexecuted_blocks=1 00:07:02.661 00:07:02.661 ' 00:07:02.661 15:12:52 spdkcli_tcp -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:07:02.661 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:02.661 --rc genhtml_branch_coverage=1 00:07:02.661 --rc genhtml_function_coverage=1 00:07:02.661 --rc genhtml_legend=1 00:07:02.661 --rc geninfo_all_blocks=1 00:07:02.661 --rc geninfo_unexecuted_blocks=1 00:07:02.661 00:07:02.661 ' 00:07:02.661 15:12:52 spdkcli_tcp -- spdkcli/tcp.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/spdkcli/common.sh 00:07:02.661 15:12:52 spdkcli_tcp -- spdkcli/common.sh@6 -- # spdkcli_job=/home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py 00:07:02.661 15:12:52 spdkcli_tcp -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/home/vagrant/spdk_repo/spdk/test/json_config/clear_config.py 00:07:02.661 15:12:52 spdkcli_tcp -- spdkcli/tcp.sh@18 -- # IP_ADDRESS=127.0.0.1 00:07:02.661 15:12:52 spdkcli_tcp -- spdkcli/tcp.sh@19 -- # PORT=9998 00:07:02.661 15:12:52 spdkcli_tcp -- spdkcli/tcp.sh@21 -- # trap 'err_cleanup; exit 1' SIGINT SIGTERM EXIT 00:07:02.661 15:12:52 spdkcli_tcp -- spdkcli/tcp.sh@23 -- # timing_enter run_spdk_tgt_tcp 00:07:02.661 15:12:52 spdkcli_tcp -- common/autotest_common.sh@726 -- # xtrace_disable 00:07:02.661 15:12:52 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:07:02.661 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:02.661 15:12:52 spdkcli_tcp -- spdkcli/tcp.sh@25 -- # spdk_tgt_pid=69824 00:07:02.661 15:12:52 spdkcli_tcp -- spdkcli/tcp.sh@27 -- # waitforlisten 69824 00:07:02.661 15:12:52 spdkcli_tcp -- common/autotest_common.sh@835 -- # '[' -z 69824 ']' 00:07:02.661 15:12:52 spdkcli_tcp -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:02.661 15:12:52 spdkcli_tcp -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:02.661 15:12:52 spdkcli_tcp -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:02.661 15:12:52 spdkcli_tcp -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:02.661 15:12:52 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:07:02.661 15:12:52 spdkcli_tcp -- spdkcli/tcp.sh@24 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x3 -p 0 00:07:02.921 [2024-11-19 15:12:53.051063] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:07:02.921 [2024-11-19 15:12:53.051207] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69824 ] 00:07:02.921 [2024-11-19 15:12:53.202263] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:07:02.921 [2024-11-19 15:12:53.247499] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:02.921 [2024-11-19 15:12:53.247603] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:07:03.862 15:12:53 spdkcli_tcp -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:03.862 15:12:53 spdkcli_tcp -- common/autotest_common.sh@868 -- # return 0 00:07:03.862 15:12:53 spdkcli_tcp -- spdkcli/tcp.sh@31 -- # socat_pid=69831 00:07:03.862 15:12:53 spdkcli_tcp -- spdkcli/tcp.sh@30 -- # socat TCP-LISTEN:9998 UNIX-CONNECT:/var/tmp/spdk.sock 00:07:03.862 15:12:53 spdkcli_tcp -- spdkcli/tcp.sh@33 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -r 100 -t 2 -s 127.0.0.1 -p 9998 rpc_get_methods 00:07:03.862 [ 00:07:03.862 "bdev_malloc_delete", 00:07:03.862 "bdev_malloc_create", 00:07:03.862 "bdev_null_resize", 00:07:03.862 "bdev_null_delete", 00:07:03.862 "bdev_null_create", 00:07:03.862 "bdev_nvme_cuse_unregister", 00:07:03.862 "bdev_nvme_cuse_register", 00:07:03.862 "bdev_opal_new_user", 00:07:03.862 "bdev_opal_set_lock_state", 00:07:03.862 "bdev_opal_delete", 00:07:03.862 "bdev_opal_get_info", 00:07:03.862 "bdev_opal_create", 00:07:03.862 "bdev_nvme_opal_revert", 00:07:03.862 "bdev_nvme_opal_init", 00:07:03.862 "bdev_nvme_send_cmd", 00:07:03.862 "bdev_nvme_set_keys", 00:07:03.862 "bdev_nvme_get_path_iostat", 00:07:03.862 "bdev_nvme_get_mdns_discovery_info", 00:07:03.862 "bdev_nvme_stop_mdns_discovery", 00:07:03.862 "bdev_nvme_start_mdns_discovery", 00:07:03.862 "bdev_nvme_set_multipath_policy", 00:07:03.862 "bdev_nvme_set_preferred_path", 00:07:03.862 "bdev_nvme_get_io_paths", 00:07:03.862 "bdev_nvme_remove_error_injection", 00:07:03.862 "bdev_nvme_add_error_injection", 00:07:03.862 "bdev_nvme_get_discovery_info", 00:07:03.862 "bdev_nvme_stop_discovery", 00:07:03.862 "bdev_nvme_start_discovery", 00:07:03.862 "bdev_nvme_get_controller_health_info", 00:07:03.862 "bdev_nvme_disable_controller", 00:07:03.862 "bdev_nvme_enable_controller", 00:07:03.862 "bdev_nvme_reset_controller", 00:07:03.862 "bdev_nvme_get_transport_statistics", 00:07:03.862 "bdev_nvme_apply_firmware", 00:07:03.862 "bdev_nvme_detach_controller", 00:07:03.862 "bdev_nvme_get_controllers", 00:07:03.862 "bdev_nvme_attach_controller", 00:07:03.862 "bdev_nvme_set_hotplug", 00:07:03.862 "bdev_nvme_set_options", 00:07:03.862 "bdev_passthru_delete", 00:07:03.862 "bdev_passthru_create", 00:07:03.862 "bdev_lvol_set_parent_bdev", 00:07:03.862 "bdev_lvol_set_parent", 00:07:03.862 "bdev_lvol_check_shallow_copy", 00:07:03.862 "bdev_lvol_start_shallow_copy", 00:07:03.862 "bdev_lvol_grow_lvstore", 00:07:03.862 "bdev_lvol_get_lvols", 00:07:03.862 "bdev_lvol_get_lvstores", 00:07:03.863 "bdev_lvol_delete", 00:07:03.863 "bdev_lvol_set_read_only", 00:07:03.863 "bdev_lvol_resize", 00:07:03.863 "bdev_lvol_decouple_parent", 00:07:03.863 "bdev_lvol_inflate", 00:07:03.863 "bdev_lvol_rename", 00:07:03.863 "bdev_lvol_clone_bdev", 00:07:03.863 "bdev_lvol_clone", 00:07:03.863 "bdev_lvol_snapshot", 00:07:03.863 "bdev_lvol_create", 00:07:03.863 "bdev_lvol_delete_lvstore", 00:07:03.863 "bdev_lvol_rename_lvstore", 00:07:03.863 "bdev_lvol_create_lvstore", 00:07:03.863 "bdev_raid_set_options", 00:07:03.863 "bdev_raid_remove_base_bdev", 00:07:03.863 "bdev_raid_add_base_bdev", 00:07:03.863 "bdev_raid_delete", 00:07:03.863 "bdev_raid_create", 00:07:03.863 "bdev_raid_get_bdevs", 00:07:03.863 "bdev_error_inject_error", 00:07:03.863 "bdev_error_delete", 00:07:03.863 "bdev_error_create", 00:07:03.863 "bdev_split_delete", 00:07:03.863 "bdev_split_create", 00:07:03.863 "bdev_delay_delete", 00:07:03.863 "bdev_delay_create", 00:07:03.863 "bdev_delay_update_latency", 00:07:03.863 "bdev_zone_block_delete", 00:07:03.863 "bdev_zone_block_create", 00:07:03.863 "blobfs_create", 00:07:03.863 "blobfs_detect", 00:07:03.863 "blobfs_set_cache_size", 00:07:03.863 "bdev_aio_delete", 00:07:03.863 "bdev_aio_rescan", 00:07:03.863 "bdev_aio_create", 00:07:03.863 "bdev_ftl_set_property", 00:07:03.863 "bdev_ftl_get_properties", 00:07:03.863 "bdev_ftl_get_stats", 00:07:03.863 "bdev_ftl_unmap", 00:07:03.863 "bdev_ftl_unload", 00:07:03.863 "bdev_ftl_delete", 00:07:03.863 "bdev_ftl_load", 00:07:03.863 "bdev_ftl_create", 00:07:03.863 "bdev_virtio_attach_controller", 00:07:03.863 "bdev_virtio_scsi_get_devices", 00:07:03.863 "bdev_virtio_detach_controller", 00:07:03.863 "bdev_virtio_blk_set_hotplug", 00:07:03.863 "bdev_iscsi_delete", 00:07:03.863 "bdev_iscsi_create", 00:07:03.863 "bdev_iscsi_set_options", 00:07:03.863 "accel_error_inject_error", 00:07:03.863 "ioat_scan_accel_module", 00:07:03.863 "dsa_scan_accel_module", 00:07:03.863 "iaa_scan_accel_module", 00:07:03.863 "keyring_file_remove_key", 00:07:03.863 "keyring_file_add_key", 00:07:03.863 "keyring_linux_set_options", 00:07:03.863 "fsdev_aio_delete", 00:07:03.863 "fsdev_aio_create", 00:07:03.863 "iscsi_get_histogram", 00:07:03.863 "iscsi_enable_histogram", 00:07:03.863 "iscsi_set_options", 00:07:03.863 "iscsi_get_auth_groups", 00:07:03.863 "iscsi_auth_group_remove_secret", 00:07:03.863 "iscsi_auth_group_add_secret", 00:07:03.863 "iscsi_delete_auth_group", 00:07:03.863 "iscsi_create_auth_group", 00:07:03.863 "iscsi_set_discovery_auth", 00:07:03.863 "iscsi_get_options", 00:07:03.863 "iscsi_target_node_request_logout", 00:07:03.863 "iscsi_target_node_set_redirect", 00:07:03.863 "iscsi_target_node_set_auth", 00:07:03.863 "iscsi_target_node_add_lun", 00:07:03.863 "iscsi_get_stats", 00:07:03.863 "iscsi_get_connections", 00:07:03.863 "iscsi_portal_group_set_auth", 00:07:03.863 "iscsi_start_portal_group", 00:07:03.863 "iscsi_delete_portal_group", 00:07:03.863 "iscsi_create_portal_group", 00:07:03.863 "iscsi_get_portal_groups", 00:07:03.863 "iscsi_delete_target_node", 00:07:03.863 "iscsi_target_node_remove_pg_ig_maps", 00:07:03.863 "iscsi_target_node_add_pg_ig_maps", 00:07:03.863 "iscsi_create_target_node", 00:07:03.863 "iscsi_get_target_nodes", 00:07:03.863 "iscsi_delete_initiator_group", 00:07:03.863 "iscsi_initiator_group_remove_initiators", 00:07:03.863 "iscsi_initiator_group_add_initiators", 00:07:03.863 "iscsi_create_initiator_group", 00:07:03.863 "iscsi_get_initiator_groups", 00:07:03.863 "nvmf_set_crdt", 00:07:03.863 "nvmf_set_config", 00:07:03.863 "nvmf_set_max_subsystems", 00:07:03.863 "nvmf_stop_mdns_prr", 00:07:03.863 "nvmf_publish_mdns_prr", 00:07:03.863 "nvmf_subsystem_get_listeners", 00:07:03.863 "nvmf_subsystem_get_qpairs", 00:07:03.863 "nvmf_subsystem_get_controllers", 00:07:03.863 "nvmf_get_stats", 00:07:03.863 "nvmf_get_transports", 00:07:03.863 "nvmf_create_transport", 00:07:03.863 "nvmf_get_targets", 00:07:03.863 "nvmf_delete_target", 00:07:03.863 "nvmf_create_target", 00:07:03.863 "nvmf_subsystem_allow_any_host", 00:07:03.863 "nvmf_subsystem_set_keys", 00:07:03.863 "nvmf_subsystem_remove_host", 00:07:03.863 "nvmf_subsystem_add_host", 00:07:03.863 "nvmf_ns_remove_host", 00:07:03.863 "nvmf_ns_add_host", 00:07:03.863 "nvmf_subsystem_remove_ns", 00:07:03.863 "nvmf_subsystem_set_ns_ana_group", 00:07:03.863 "nvmf_subsystem_add_ns", 00:07:03.863 "nvmf_subsystem_listener_set_ana_state", 00:07:03.863 "nvmf_discovery_get_referrals", 00:07:03.863 "nvmf_discovery_remove_referral", 00:07:03.863 "nvmf_discovery_add_referral", 00:07:03.863 "nvmf_subsystem_remove_listener", 00:07:03.863 "nvmf_subsystem_add_listener", 00:07:03.863 "nvmf_delete_subsystem", 00:07:03.863 "nvmf_create_subsystem", 00:07:03.863 "nvmf_get_subsystems", 00:07:03.863 "env_dpdk_get_mem_stats", 00:07:03.863 "nbd_get_disks", 00:07:03.863 "nbd_stop_disk", 00:07:03.863 "nbd_start_disk", 00:07:03.863 "ublk_recover_disk", 00:07:03.863 "ublk_get_disks", 00:07:03.863 "ublk_stop_disk", 00:07:03.863 "ublk_start_disk", 00:07:03.863 "ublk_destroy_target", 00:07:03.863 "ublk_create_target", 00:07:03.863 "virtio_blk_create_transport", 00:07:03.863 "virtio_blk_get_transports", 00:07:03.863 "vhost_controller_set_coalescing", 00:07:03.863 "vhost_get_controllers", 00:07:03.863 "vhost_delete_controller", 00:07:03.863 "vhost_create_blk_controller", 00:07:03.863 "vhost_scsi_controller_remove_target", 00:07:03.863 "vhost_scsi_controller_add_target", 00:07:03.863 "vhost_start_scsi_controller", 00:07:03.863 "vhost_create_scsi_controller", 00:07:03.863 "thread_set_cpumask", 00:07:03.863 "scheduler_set_options", 00:07:03.863 "framework_get_governor", 00:07:03.863 "framework_get_scheduler", 00:07:03.863 "framework_set_scheduler", 00:07:03.863 "framework_get_reactors", 00:07:03.863 "thread_get_io_channels", 00:07:03.863 "thread_get_pollers", 00:07:03.863 "thread_get_stats", 00:07:03.863 "framework_monitor_context_switch", 00:07:03.863 "spdk_kill_instance", 00:07:03.863 "log_enable_timestamps", 00:07:03.863 "log_get_flags", 00:07:03.863 "log_clear_flag", 00:07:03.863 "log_set_flag", 00:07:03.863 "log_get_level", 00:07:03.863 "log_set_level", 00:07:03.863 "log_get_print_level", 00:07:03.863 "log_set_print_level", 00:07:03.863 "framework_enable_cpumask_locks", 00:07:03.863 "framework_disable_cpumask_locks", 00:07:03.863 "framework_wait_init", 00:07:03.863 "framework_start_init", 00:07:03.863 "scsi_get_devices", 00:07:03.863 "bdev_get_histogram", 00:07:03.863 "bdev_enable_histogram", 00:07:03.863 "bdev_set_qos_limit", 00:07:03.863 "bdev_set_qd_sampling_period", 00:07:03.863 "bdev_get_bdevs", 00:07:03.863 "bdev_reset_iostat", 00:07:03.863 "bdev_get_iostat", 00:07:03.863 "bdev_examine", 00:07:03.863 "bdev_wait_for_examine", 00:07:03.863 "bdev_set_options", 00:07:03.863 "accel_get_stats", 00:07:03.863 "accel_set_options", 00:07:03.863 "accel_set_driver", 00:07:03.863 "accel_crypto_key_destroy", 00:07:03.863 "accel_crypto_keys_get", 00:07:03.863 "accel_crypto_key_create", 00:07:03.863 "accel_assign_opc", 00:07:03.863 "accel_get_module_info", 00:07:03.863 "accel_get_opc_assignments", 00:07:03.863 "vmd_rescan", 00:07:03.863 "vmd_remove_device", 00:07:03.863 "vmd_enable", 00:07:03.863 "sock_get_default_impl", 00:07:03.863 "sock_set_default_impl", 00:07:03.863 "sock_impl_set_options", 00:07:03.863 "sock_impl_get_options", 00:07:03.863 "iobuf_get_stats", 00:07:03.863 "iobuf_set_options", 00:07:03.863 "keyring_get_keys", 00:07:03.863 "framework_get_pci_devices", 00:07:03.863 "framework_get_config", 00:07:03.863 "framework_get_subsystems", 00:07:03.863 "fsdev_set_opts", 00:07:03.863 "fsdev_get_opts", 00:07:03.863 "trace_get_info", 00:07:03.863 "trace_get_tpoint_group_mask", 00:07:03.863 "trace_disable_tpoint_group", 00:07:03.863 "trace_enable_tpoint_group", 00:07:03.863 "trace_clear_tpoint_mask", 00:07:03.863 "trace_set_tpoint_mask", 00:07:03.863 "notify_get_notifications", 00:07:03.863 "notify_get_types", 00:07:03.863 "spdk_get_version", 00:07:03.863 "rpc_get_methods" 00:07:03.863 ] 00:07:03.863 15:12:54 spdkcli_tcp -- spdkcli/tcp.sh@35 -- # timing_exit run_spdk_tgt_tcp 00:07:03.863 15:12:54 spdkcli_tcp -- common/autotest_common.sh@732 -- # xtrace_disable 00:07:03.863 15:12:54 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:07:03.863 15:12:54 spdkcli_tcp -- spdkcli/tcp.sh@37 -- # trap - SIGINT SIGTERM EXIT 00:07:03.863 15:12:54 spdkcli_tcp -- spdkcli/tcp.sh@38 -- # killprocess 69824 00:07:03.863 15:12:54 spdkcli_tcp -- common/autotest_common.sh@954 -- # '[' -z 69824 ']' 00:07:03.863 15:12:54 spdkcli_tcp -- common/autotest_common.sh@958 -- # kill -0 69824 00:07:03.863 15:12:54 spdkcli_tcp -- common/autotest_common.sh@959 -- # uname 00:07:03.863 15:12:54 spdkcli_tcp -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:03.863 15:12:54 spdkcli_tcp -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 69824 00:07:03.863 15:12:54 spdkcli_tcp -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:03.863 15:12:54 spdkcli_tcp -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:03.863 15:12:54 spdkcli_tcp -- common/autotest_common.sh@972 -- # echo 'killing process with pid 69824' 00:07:03.863 killing process with pid 69824 00:07:03.863 15:12:54 spdkcli_tcp -- common/autotest_common.sh@973 -- # kill 69824 00:07:03.863 15:12:54 spdkcli_tcp -- common/autotest_common.sh@978 -- # wait 69824 00:07:04.804 00:07:04.804 real 0m2.050s 00:07:04.804 user 0m3.383s 00:07:04.804 sys 0m0.702s 00:07:04.804 ************************************ 00:07:04.804 END TEST spdkcli_tcp 00:07:04.804 ************************************ 00:07:04.804 15:12:54 spdkcli_tcp -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:04.804 15:12:54 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:07:04.804 15:12:54 -- spdk/autotest.sh@167 -- # run_test dpdk_mem_utility /home/vagrant/spdk_repo/spdk/test/dpdk_memory_utility/test_dpdk_mem_info.sh 00:07:04.804 15:12:54 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:07:04.804 15:12:54 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:04.804 15:12:54 -- common/autotest_common.sh@10 -- # set +x 00:07:04.804 ************************************ 00:07:04.804 START TEST dpdk_mem_utility 00:07:04.804 ************************************ 00:07:04.804 15:12:54 dpdk_mem_utility -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/dpdk_memory_utility/test_dpdk_mem_info.sh 00:07:04.804 * Looking for test storage... 00:07:04.804 * Found test storage at /home/vagrant/spdk_repo/spdk/test/dpdk_memory_utility 00:07:04.804 15:12:54 dpdk_mem_utility -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:07:04.804 15:12:54 dpdk_mem_utility -- common/autotest_common.sh@1693 -- # lcov --version 00:07:04.804 15:12:54 dpdk_mem_utility -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:07:04.804 15:12:55 dpdk_mem_utility -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:07:04.804 15:12:55 dpdk_mem_utility -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:07:04.804 15:12:55 dpdk_mem_utility -- scripts/common.sh@333 -- # local ver1 ver1_l 00:07:04.804 15:12:55 dpdk_mem_utility -- scripts/common.sh@334 -- # local ver2 ver2_l 00:07:04.804 15:12:55 dpdk_mem_utility -- scripts/common.sh@336 -- # IFS=.-: 00:07:04.804 15:12:55 dpdk_mem_utility -- scripts/common.sh@336 -- # read -ra ver1 00:07:04.804 15:12:55 dpdk_mem_utility -- scripts/common.sh@337 -- # IFS=.-: 00:07:04.804 15:12:55 dpdk_mem_utility -- scripts/common.sh@337 -- # read -ra ver2 00:07:04.804 15:12:55 dpdk_mem_utility -- scripts/common.sh@338 -- # local 'op=<' 00:07:04.804 15:12:55 dpdk_mem_utility -- scripts/common.sh@340 -- # ver1_l=2 00:07:04.804 15:12:55 dpdk_mem_utility -- scripts/common.sh@341 -- # ver2_l=1 00:07:04.804 15:12:55 dpdk_mem_utility -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:07:04.804 15:12:55 dpdk_mem_utility -- scripts/common.sh@344 -- # case "$op" in 00:07:04.804 15:12:55 dpdk_mem_utility -- scripts/common.sh@345 -- # : 1 00:07:04.804 15:12:55 dpdk_mem_utility -- scripts/common.sh@364 -- # (( v = 0 )) 00:07:04.804 15:12:55 dpdk_mem_utility -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:07:04.804 15:12:55 dpdk_mem_utility -- scripts/common.sh@365 -- # decimal 1 00:07:04.804 15:12:55 dpdk_mem_utility -- scripts/common.sh@353 -- # local d=1 00:07:04.804 15:12:55 dpdk_mem_utility -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:07:04.804 15:12:55 dpdk_mem_utility -- scripts/common.sh@355 -- # echo 1 00:07:04.804 15:12:55 dpdk_mem_utility -- scripts/common.sh@365 -- # ver1[v]=1 00:07:04.804 15:12:55 dpdk_mem_utility -- scripts/common.sh@366 -- # decimal 2 00:07:04.804 15:12:55 dpdk_mem_utility -- scripts/common.sh@353 -- # local d=2 00:07:04.804 15:12:55 dpdk_mem_utility -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:07:04.804 15:12:55 dpdk_mem_utility -- scripts/common.sh@355 -- # echo 2 00:07:04.804 15:12:55 dpdk_mem_utility -- scripts/common.sh@366 -- # ver2[v]=2 00:07:04.804 15:12:55 dpdk_mem_utility -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:07:04.804 15:12:55 dpdk_mem_utility -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:07:04.804 15:12:55 dpdk_mem_utility -- scripts/common.sh@368 -- # return 0 00:07:04.804 15:12:55 dpdk_mem_utility -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:07:04.804 15:12:55 dpdk_mem_utility -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:07:04.804 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:04.804 --rc genhtml_branch_coverage=1 00:07:04.804 --rc genhtml_function_coverage=1 00:07:04.804 --rc genhtml_legend=1 00:07:04.804 --rc geninfo_all_blocks=1 00:07:04.804 --rc geninfo_unexecuted_blocks=1 00:07:04.804 00:07:04.804 ' 00:07:04.804 15:12:55 dpdk_mem_utility -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:07:04.804 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:04.804 --rc genhtml_branch_coverage=1 00:07:04.804 --rc genhtml_function_coverage=1 00:07:04.804 --rc genhtml_legend=1 00:07:04.804 --rc geninfo_all_blocks=1 00:07:04.804 --rc geninfo_unexecuted_blocks=1 00:07:04.804 00:07:04.804 ' 00:07:04.804 15:12:55 dpdk_mem_utility -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:07:04.804 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:04.804 --rc genhtml_branch_coverage=1 00:07:04.804 --rc genhtml_function_coverage=1 00:07:04.804 --rc genhtml_legend=1 00:07:04.804 --rc geninfo_all_blocks=1 00:07:04.804 --rc geninfo_unexecuted_blocks=1 00:07:04.804 00:07:04.804 ' 00:07:04.804 15:12:55 dpdk_mem_utility -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:07:04.804 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:04.804 --rc genhtml_branch_coverage=1 00:07:04.804 --rc genhtml_function_coverage=1 00:07:04.804 --rc genhtml_legend=1 00:07:04.804 --rc geninfo_all_blocks=1 00:07:04.804 --rc geninfo_unexecuted_blocks=1 00:07:04.804 00:07:04.804 ' 00:07:04.804 15:12:55 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@10 -- # MEM_SCRIPT=/home/vagrant/spdk_repo/spdk/scripts/dpdk_mem_info.py 00:07:04.804 15:12:55 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@13 -- # spdkpid=69925 00:07:04.804 15:12:55 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@12 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:07:04.804 15:12:55 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@15 -- # waitforlisten 69925 00:07:04.805 15:12:55 dpdk_mem_utility -- common/autotest_common.sh@835 -- # '[' -z 69925 ']' 00:07:04.805 15:12:55 dpdk_mem_utility -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:04.805 15:12:55 dpdk_mem_utility -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:04.805 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:04.805 15:12:55 dpdk_mem_utility -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:04.805 15:12:55 dpdk_mem_utility -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:04.805 15:12:55 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:07:05.064 [2024-11-19 15:12:55.205063] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:07:05.064 [2024-11-19 15:12:55.205362] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69925 ] 00:07:05.064 [2024-11-19 15:12:55.361911] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:05.324 [2024-11-19 15:12:55.402458] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:05.895 15:12:56 dpdk_mem_utility -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:05.895 15:12:56 dpdk_mem_utility -- common/autotest_common.sh@868 -- # return 0 00:07:05.895 15:12:56 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@17 -- # trap 'killprocess $spdkpid' SIGINT SIGTERM EXIT 00:07:05.895 15:12:56 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@19 -- # rpc_cmd env_dpdk_get_mem_stats 00:07:05.895 15:12:56 dpdk_mem_utility -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:05.895 15:12:56 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:07:05.895 { 00:07:05.895 "filename": "/tmp/spdk_mem_dump.txt" 00:07:05.895 } 00:07:05.895 15:12:56 dpdk_mem_utility -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:05.895 15:12:56 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@21 -- # /home/vagrant/spdk_repo/spdk/scripts/dpdk_mem_info.py 00:07:05.895 DPDK memory size 810.000000 MiB in 1 heap(s) 00:07:05.895 1 heaps totaling size 810.000000 MiB 00:07:05.895 size: 810.000000 MiB heap id: 0 00:07:05.895 end heaps---------- 00:07:05.895 9 mempools totaling size 595.772034 MiB 00:07:05.895 size: 212.674988 MiB name: PDU_immediate_data_Pool 00:07:05.895 size: 158.602051 MiB name: PDU_data_out_Pool 00:07:05.895 size: 92.545471 MiB name: bdev_io_69925 00:07:05.895 size: 50.003479 MiB name: msgpool_69925 00:07:05.895 size: 36.509338 MiB name: fsdev_io_69925 00:07:05.895 size: 21.763794 MiB name: PDU_Pool 00:07:05.895 size: 19.513306 MiB name: SCSI_TASK_Pool 00:07:05.895 size: 4.133484 MiB name: evtpool_69925 00:07:05.895 size: 0.026123 MiB name: Session_Pool 00:07:05.895 end mempools------- 00:07:05.895 6 memzones totaling size 4.142822 MiB 00:07:05.895 size: 1.000366 MiB name: RG_ring_0_69925 00:07:05.895 size: 1.000366 MiB name: RG_ring_1_69925 00:07:05.895 size: 1.000366 MiB name: RG_ring_4_69925 00:07:05.895 size: 1.000366 MiB name: RG_ring_5_69925 00:07:05.895 size: 0.125366 MiB name: RG_ring_2_69925 00:07:05.895 size: 0.015991 MiB name: RG_ring_3_69925 00:07:05.895 end memzones------- 00:07:05.895 15:12:56 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@23 -- # /home/vagrant/spdk_repo/spdk/scripts/dpdk_mem_info.py -m 0 00:07:05.895 heap id: 0 total size: 810.000000 MiB number of busy elements: 311 number of free elements: 15 00:07:05.895 list of free elements. size: 10.813599 MiB 00:07:05.895 element at address: 0x200018a00000 with size: 0.999878 MiB 00:07:05.895 element at address: 0x200018c00000 with size: 0.999878 MiB 00:07:05.895 element at address: 0x200031800000 with size: 0.994446 MiB 00:07:05.895 element at address: 0x200000400000 with size: 0.993958 MiB 00:07:05.895 element at address: 0x200006400000 with size: 0.959839 MiB 00:07:05.895 element at address: 0x200012c00000 with size: 0.954285 MiB 00:07:05.895 element at address: 0x200018e00000 with size: 0.936584 MiB 00:07:05.895 element at address: 0x200000200000 with size: 0.717346 MiB 00:07:05.895 element at address: 0x20001a600000 with size: 0.567505 MiB 00:07:05.895 element at address: 0x20000a600000 with size: 0.488892 MiB 00:07:05.895 element at address: 0x200000c00000 with size: 0.487000 MiB 00:07:05.895 element at address: 0x200019000000 with size: 0.485657 MiB 00:07:05.895 element at address: 0x200003e00000 with size: 0.480286 MiB 00:07:05.895 element at address: 0x200027a00000 with size: 0.396301 MiB 00:07:05.895 element at address: 0x200000800000 with size: 0.351746 MiB 00:07:05.895 list of standard malloc elements. size: 199.267517 MiB 00:07:05.895 element at address: 0x20000a7fff80 with size: 132.000122 MiB 00:07:05.895 element at address: 0x2000065fff80 with size: 64.000122 MiB 00:07:05.895 element at address: 0x200018afff80 with size: 1.000122 MiB 00:07:05.895 element at address: 0x200018cfff80 with size: 1.000122 MiB 00:07:05.895 element at address: 0x200018efff80 with size: 1.000122 MiB 00:07:05.895 element at address: 0x2000003d9f00 with size: 0.140747 MiB 00:07:05.895 element at address: 0x200018eeff00 with size: 0.062622 MiB 00:07:05.895 element at address: 0x2000003fdf80 with size: 0.007935 MiB 00:07:05.895 element at address: 0x200018eefdc0 with size: 0.000305 MiB 00:07:05.895 element at address: 0x2000002d7c40 with size: 0.000183 MiB 00:07:05.895 element at address: 0x2000003d9e40 with size: 0.000183 MiB 00:07:05.895 element at address: 0x2000004fe740 with size: 0.000183 MiB 00:07:05.895 element at address: 0x2000004fe800 with size: 0.000183 MiB 00:07:05.895 element at address: 0x2000004fe8c0 with size: 0.000183 MiB 00:07:05.895 element at address: 0x2000004fe980 with size: 0.000183 MiB 00:07:05.895 element at address: 0x2000004fea40 with size: 0.000183 MiB 00:07:05.895 element at address: 0x2000004feb00 with size: 0.000183 MiB 00:07:05.895 element at address: 0x2000004febc0 with size: 0.000183 MiB 00:07:05.895 element at address: 0x2000004fec80 with size: 0.000183 MiB 00:07:05.895 element at address: 0x2000004fed40 with size: 0.000183 MiB 00:07:05.895 element at address: 0x2000004fee00 with size: 0.000183 MiB 00:07:05.895 element at address: 0x2000004feec0 with size: 0.000183 MiB 00:07:05.895 element at address: 0x2000004fef80 with size: 0.000183 MiB 00:07:05.895 element at address: 0x2000004ff040 with size: 0.000183 MiB 00:07:05.895 element at address: 0x2000004ff100 with size: 0.000183 MiB 00:07:05.895 element at address: 0x2000004ff1c0 with size: 0.000183 MiB 00:07:05.895 element at address: 0x2000004ff280 with size: 0.000183 MiB 00:07:05.895 element at address: 0x2000004ff340 with size: 0.000183 MiB 00:07:05.895 element at address: 0x2000004ff400 with size: 0.000183 MiB 00:07:05.895 element at address: 0x2000004ff4c0 with size: 0.000183 MiB 00:07:05.895 element at address: 0x2000004ff580 with size: 0.000183 MiB 00:07:05.895 element at address: 0x2000004ff640 with size: 0.000183 MiB 00:07:05.895 element at address: 0x2000004ff700 with size: 0.000183 MiB 00:07:05.895 element at address: 0x2000004ff7c0 with size: 0.000183 MiB 00:07:05.895 element at address: 0x2000004ff880 with size: 0.000183 MiB 00:07:05.895 element at address: 0x2000004ff940 with size: 0.000183 MiB 00:07:05.895 element at address: 0x2000004ffa00 with size: 0.000183 MiB 00:07:05.895 element at address: 0x2000004ffac0 with size: 0.000183 MiB 00:07:05.895 element at address: 0x2000004ffcc0 with size: 0.000183 MiB 00:07:05.895 element at address: 0x2000004ffd80 with size: 0.000183 MiB 00:07:05.895 element at address: 0x2000004ffe40 with size: 0.000183 MiB 00:07:05.895 element at address: 0x20000085a0c0 with size: 0.000183 MiB 00:07:05.895 element at address: 0x20000085a2c0 with size: 0.000183 MiB 00:07:05.895 element at address: 0x20000085e580 with size: 0.000183 MiB 00:07:05.895 element at address: 0x20000087e840 with size: 0.000183 MiB 00:07:05.895 element at address: 0x20000087e900 with size: 0.000183 MiB 00:07:05.895 element at address: 0x20000087e9c0 with size: 0.000183 MiB 00:07:05.895 element at address: 0x20000087ea80 with size: 0.000183 MiB 00:07:05.895 element at address: 0x20000087eb40 with size: 0.000183 MiB 00:07:05.895 element at address: 0x20000087ec00 with size: 0.000183 MiB 00:07:05.895 element at address: 0x20000087ecc0 with size: 0.000183 MiB 00:07:05.895 element at address: 0x20000087ed80 with size: 0.000183 MiB 00:07:05.895 element at address: 0x20000087ee40 with size: 0.000183 MiB 00:07:05.895 element at address: 0x20000087ef00 with size: 0.000183 MiB 00:07:05.895 element at address: 0x20000087efc0 with size: 0.000183 MiB 00:07:05.895 element at address: 0x20000087f080 with size: 0.000183 MiB 00:07:05.895 element at address: 0x20000087f140 with size: 0.000183 MiB 00:07:05.895 element at address: 0x20000087f200 with size: 0.000183 MiB 00:07:05.895 element at address: 0x20000087f2c0 with size: 0.000183 MiB 00:07:05.895 element at address: 0x20000087f380 with size: 0.000183 MiB 00:07:05.895 element at address: 0x20000087f440 with size: 0.000183 MiB 00:07:05.895 element at address: 0x20000087f500 with size: 0.000183 MiB 00:07:05.895 element at address: 0x20000087f5c0 with size: 0.000183 MiB 00:07:05.895 element at address: 0x20000087f680 with size: 0.000183 MiB 00:07:05.896 element at address: 0x2000008ff940 with size: 0.000183 MiB 00:07:05.896 element at address: 0x2000008ffb40 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7cac0 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7cb80 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7cc40 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7cd00 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7cdc0 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7ce80 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7cf40 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7d000 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7d0c0 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7d180 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7d240 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7d300 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7d3c0 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7d480 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7d540 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7d600 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7d6c0 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7d780 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7d840 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7d900 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7d9c0 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7da80 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7db40 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7dc00 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7dcc0 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7dd80 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7de40 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7df00 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7dfc0 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7e080 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7e140 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7e200 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7e2c0 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7e380 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7e440 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7e500 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7e5c0 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7e680 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7e740 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7e800 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7e8c0 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7e980 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7ea40 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7eb00 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7ebc0 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7ec80 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000c7ed40 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000cff000 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200000cff0c0 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200003e7af40 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200003e7b000 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200003e7b0c0 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200003e7b180 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200003e7b240 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200003e7b300 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200003e7b3c0 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200003e7b480 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200003e7b540 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200003e7b600 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200003e7b6c0 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200003efb980 with size: 0.000183 MiB 00:07:05.896 element at address: 0x2000064fdd80 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20000a67d280 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20000a67d340 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20000a67d400 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20000a67d4c0 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20000a67d580 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20000a67d640 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20000a67d700 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20000a67d7c0 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20000a67d880 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20000a67d940 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20000a67da00 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20000a67dac0 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20000a6fdd80 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200012cf44c0 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200018eefc40 with size: 0.000183 MiB 00:07:05.896 element at address: 0x200018eefd00 with size: 0.000183 MiB 00:07:05.896 element at address: 0x2000190bc740 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a691480 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a691540 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a691600 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a6916c0 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a691780 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a691840 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a691900 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a6919c0 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a691a80 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a691b40 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a691c00 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a691cc0 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a691d80 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a691e40 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a691f00 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a691fc0 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a692080 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a692140 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a692200 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a6922c0 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a692380 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a692440 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a692500 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a6925c0 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a692680 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a692740 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a692800 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a6928c0 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a692980 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a692a40 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a692b00 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a692bc0 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a692c80 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a692d40 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a692e00 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a692ec0 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a692f80 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a693040 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a693100 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a6931c0 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a693280 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a693340 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a693400 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a6934c0 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a693580 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a693640 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a693700 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a6937c0 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a693880 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a693940 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a693a00 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a693ac0 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a693b80 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a693c40 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a693d00 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a693dc0 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a693e80 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a693f40 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a694000 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a6940c0 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a694180 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a694240 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a694300 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a6943c0 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a694480 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a694540 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a694600 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a6946c0 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a694780 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a694840 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a694900 with size: 0.000183 MiB 00:07:05.896 element at address: 0x20001a6949c0 with size: 0.000183 MiB 00:07:05.897 element at address: 0x20001a694a80 with size: 0.000183 MiB 00:07:05.897 element at address: 0x20001a694b40 with size: 0.000183 MiB 00:07:05.897 element at address: 0x20001a694c00 with size: 0.000183 MiB 00:07:05.897 element at address: 0x20001a694cc0 with size: 0.000183 MiB 00:07:05.897 element at address: 0x20001a694d80 with size: 0.000183 MiB 00:07:05.897 element at address: 0x20001a694e40 with size: 0.000183 MiB 00:07:05.897 element at address: 0x20001a694f00 with size: 0.000183 MiB 00:07:05.897 element at address: 0x20001a694fc0 with size: 0.000183 MiB 00:07:05.897 element at address: 0x20001a695080 with size: 0.000183 MiB 00:07:05.897 element at address: 0x20001a695140 with size: 0.000183 MiB 00:07:05.897 element at address: 0x20001a695200 with size: 0.000183 MiB 00:07:05.897 element at address: 0x20001a6952c0 with size: 0.000183 MiB 00:07:05.897 element at address: 0x20001a695380 with size: 0.000183 MiB 00:07:05.897 element at address: 0x20001a695440 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a65740 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a65800 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6c400 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6c600 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6c6c0 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6c780 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6c840 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6c900 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6c9c0 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6ca80 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6cb40 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6cc00 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6ccc0 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6cd80 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6ce40 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6cf00 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6cfc0 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6d080 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6d140 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6d200 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6d2c0 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6d380 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6d440 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6d500 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6d5c0 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6d680 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6d740 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6d800 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6d8c0 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6d980 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6da40 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6db00 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6dbc0 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6dc80 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6dd40 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6de00 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6dec0 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6df80 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6e040 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6e100 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6e1c0 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6e280 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6e340 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6e400 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6e4c0 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6e580 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6e640 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6e700 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6e7c0 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6e880 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6e940 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6ea00 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6eac0 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6eb80 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6ec40 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6ed00 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6edc0 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6ee80 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6ef40 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6f000 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6f0c0 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6f180 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6f240 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6f300 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6f3c0 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6f480 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6f540 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6f600 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6f6c0 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6f780 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6f840 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6f900 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6f9c0 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6fa80 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6fb40 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6fc00 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6fcc0 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6fd80 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6fe40 with size: 0.000183 MiB 00:07:05.897 element at address: 0x200027a6ff00 with size: 0.000183 MiB 00:07:05.897 list of memzone associated elements. size: 599.918884 MiB 00:07:05.897 element at address: 0x20001a695500 with size: 211.416748 MiB 00:07:05.897 associated memzone info: size: 211.416626 MiB name: MP_PDU_immediate_data_Pool_0 00:07:05.897 element at address: 0x200027a6ffc0 with size: 157.562561 MiB 00:07:05.897 associated memzone info: size: 157.562439 MiB name: MP_PDU_data_out_Pool_0 00:07:05.897 element at address: 0x200012df4780 with size: 92.045044 MiB 00:07:05.897 associated memzone info: size: 92.044922 MiB name: MP_bdev_io_69925_0 00:07:05.897 element at address: 0x200000dff380 with size: 48.003052 MiB 00:07:05.897 associated memzone info: size: 48.002930 MiB name: MP_msgpool_69925_0 00:07:05.897 element at address: 0x200003ffdb80 with size: 36.008911 MiB 00:07:05.897 associated memzone info: size: 36.008789 MiB name: MP_fsdev_io_69925_0 00:07:05.897 element at address: 0x2000191be940 with size: 20.255554 MiB 00:07:05.897 associated memzone info: size: 20.255432 MiB name: MP_PDU_Pool_0 00:07:05.897 element at address: 0x2000319feb40 with size: 18.005066 MiB 00:07:05.897 associated memzone info: size: 18.004944 MiB name: MP_SCSI_TASK_Pool_0 00:07:05.897 element at address: 0x2000004fff00 with size: 3.000244 MiB 00:07:05.897 associated memzone info: size: 3.000122 MiB name: MP_evtpool_69925_0 00:07:05.897 element at address: 0x2000009ffe00 with size: 2.000488 MiB 00:07:05.897 associated memzone info: size: 2.000366 MiB name: RG_MP_msgpool_69925 00:07:05.897 element at address: 0x2000002d7d00 with size: 1.008118 MiB 00:07:05.897 associated memzone info: size: 1.007996 MiB name: MP_evtpool_69925 00:07:05.897 element at address: 0x20000a6fde40 with size: 1.008118 MiB 00:07:05.897 associated memzone info: size: 1.007996 MiB name: MP_PDU_Pool 00:07:05.897 element at address: 0x2000190bc800 with size: 1.008118 MiB 00:07:05.897 associated memzone info: size: 1.007996 MiB name: MP_PDU_immediate_data_Pool 00:07:05.897 element at address: 0x2000064fde40 with size: 1.008118 MiB 00:07:05.897 associated memzone info: size: 1.007996 MiB name: MP_PDU_data_out_Pool 00:07:05.897 element at address: 0x200003efba40 with size: 1.008118 MiB 00:07:05.897 associated memzone info: size: 1.007996 MiB name: MP_SCSI_TASK_Pool 00:07:05.897 element at address: 0x200000cff180 with size: 1.000488 MiB 00:07:05.897 associated memzone info: size: 1.000366 MiB name: RG_ring_0_69925 00:07:05.897 element at address: 0x2000008ffc00 with size: 1.000488 MiB 00:07:05.897 associated memzone info: size: 1.000366 MiB name: RG_ring_1_69925 00:07:05.897 element at address: 0x200012cf4580 with size: 1.000488 MiB 00:07:05.897 associated memzone info: size: 1.000366 MiB name: RG_ring_4_69925 00:07:05.897 element at address: 0x2000318fe940 with size: 1.000488 MiB 00:07:05.897 associated memzone info: size: 1.000366 MiB name: RG_ring_5_69925 00:07:05.897 element at address: 0x20000087f740 with size: 0.500488 MiB 00:07:05.897 associated memzone info: size: 0.500366 MiB name: RG_MP_fsdev_io_69925 00:07:05.897 element at address: 0x200000c7ee00 with size: 0.500488 MiB 00:07:05.897 associated memzone info: size: 0.500366 MiB name: RG_MP_bdev_io_69925 00:07:05.897 element at address: 0x20000a67db80 with size: 0.500488 MiB 00:07:05.897 associated memzone info: size: 0.500366 MiB name: RG_MP_PDU_Pool 00:07:05.897 element at address: 0x200003e7b780 with size: 0.500488 MiB 00:07:05.897 associated memzone info: size: 0.500366 MiB name: RG_MP_SCSI_TASK_Pool 00:07:05.897 element at address: 0x20001907c540 with size: 0.250488 MiB 00:07:05.897 associated memzone info: size: 0.250366 MiB name: RG_MP_PDU_immediate_data_Pool 00:07:05.897 element at address: 0x2000002b7a40 with size: 0.125488 MiB 00:07:05.897 associated memzone info: size: 0.125366 MiB name: RG_MP_evtpool_69925 00:07:05.897 element at address: 0x20000085e640 with size: 0.125488 MiB 00:07:05.897 associated memzone info: size: 0.125366 MiB name: RG_ring_2_69925 00:07:05.897 element at address: 0x2000064f5b80 with size: 0.031738 MiB 00:07:05.897 associated memzone info: size: 0.031616 MiB name: RG_MP_PDU_data_out_Pool 00:07:05.897 element at address: 0x200027a658c0 with size: 0.023743 MiB 00:07:05.897 associated memzone info: size: 0.023621 MiB name: MP_Session_Pool_0 00:07:05.898 element at address: 0x20000085a380 with size: 0.016113 MiB 00:07:05.898 associated memzone info: size: 0.015991 MiB name: RG_ring_3_69925 00:07:05.898 element at address: 0x200027a6ba00 with size: 0.002441 MiB 00:07:05.898 associated memzone info: size: 0.002319 MiB name: RG_MP_Session_Pool 00:07:05.898 element at address: 0x2000004ffb80 with size: 0.000305 MiB 00:07:05.898 associated memzone info: size: 0.000183 MiB name: MP_msgpool_69925 00:07:05.898 element at address: 0x2000008ffa00 with size: 0.000305 MiB 00:07:05.898 associated memzone info: size: 0.000183 MiB name: MP_fsdev_io_69925 00:07:05.898 element at address: 0x20000085a180 with size: 0.000305 MiB 00:07:05.898 associated memzone info: size: 0.000183 MiB name: MP_bdev_io_69925 00:07:05.898 element at address: 0x200027a6c4c0 with size: 0.000305 MiB 00:07:05.898 associated memzone info: size: 0.000183 MiB name: MP_Session_Pool 00:07:05.898 15:12:56 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@25 -- # trap - SIGINT SIGTERM EXIT 00:07:05.898 15:12:56 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@26 -- # killprocess 69925 00:07:05.898 15:12:56 dpdk_mem_utility -- common/autotest_common.sh@954 -- # '[' -z 69925 ']' 00:07:05.898 15:12:56 dpdk_mem_utility -- common/autotest_common.sh@958 -- # kill -0 69925 00:07:05.898 15:12:56 dpdk_mem_utility -- common/autotest_common.sh@959 -- # uname 00:07:05.898 15:12:56 dpdk_mem_utility -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:05.898 15:12:56 dpdk_mem_utility -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 69925 00:07:05.898 15:12:56 dpdk_mem_utility -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:05.898 15:12:56 dpdk_mem_utility -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:05.898 15:12:56 dpdk_mem_utility -- common/autotest_common.sh@972 -- # echo 'killing process with pid 69925' 00:07:05.898 killing process with pid 69925 00:07:05.898 15:12:56 dpdk_mem_utility -- common/autotest_common.sh@973 -- # kill 69925 00:07:05.898 15:12:56 dpdk_mem_utility -- common/autotest_common.sh@978 -- # wait 69925 00:07:06.467 00:07:06.467 real 0m1.913s 00:07:06.467 user 0m1.720s 00:07:06.467 sys 0m0.654s 00:07:06.467 15:12:56 dpdk_mem_utility -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:06.467 15:12:56 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:07:06.467 ************************************ 00:07:06.467 END TEST dpdk_mem_utility 00:07:06.467 ************************************ 00:07:06.727 15:12:56 -- spdk/autotest.sh@168 -- # run_test event /home/vagrant/spdk_repo/spdk/test/event/event.sh 00:07:06.727 15:12:56 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:07:06.727 15:12:56 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:06.727 15:12:56 -- common/autotest_common.sh@10 -- # set +x 00:07:06.727 ************************************ 00:07:06.727 START TEST event 00:07:06.727 ************************************ 00:07:06.727 15:12:56 event -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/event/event.sh 00:07:06.727 * Looking for test storage... 00:07:06.727 * Found test storage at /home/vagrant/spdk_repo/spdk/test/event 00:07:06.727 15:12:56 event -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:07:06.727 15:12:56 event -- common/autotest_common.sh@1693 -- # lcov --version 00:07:06.727 15:12:56 event -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:07:06.727 15:12:57 event -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:07:06.727 15:12:57 event -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:07:06.727 15:12:57 event -- scripts/common.sh@333 -- # local ver1 ver1_l 00:07:06.727 15:12:57 event -- scripts/common.sh@334 -- # local ver2 ver2_l 00:07:06.727 15:12:57 event -- scripts/common.sh@336 -- # IFS=.-: 00:07:06.727 15:12:57 event -- scripts/common.sh@336 -- # read -ra ver1 00:07:06.727 15:12:57 event -- scripts/common.sh@337 -- # IFS=.-: 00:07:06.727 15:12:57 event -- scripts/common.sh@337 -- # read -ra ver2 00:07:06.727 15:12:57 event -- scripts/common.sh@338 -- # local 'op=<' 00:07:06.727 15:12:57 event -- scripts/common.sh@340 -- # ver1_l=2 00:07:06.727 15:12:57 event -- scripts/common.sh@341 -- # ver2_l=1 00:07:06.727 15:12:57 event -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:07:06.727 15:12:57 event -- scripts/common.sh@344 -- # case "$op" in 00:07:06.727 15:12:57 event -- scripts/common.sh@345 -- # : 1 00:07:06.727 15:12:57 event -- scripts/common.sh@364 -- # (( v = 0 )) 00:07:06.727 15:12:57 event -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:07:06.727 15:12:57 event -- scripts/common.sh@365 -- # decimal 1 00:07:06.727 15:12:57 event -- scripts/common.sh@353 -- # local d=1 00:07:06.727 15:12:57 event -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:07:06.727 15:12:57 event -- scripts/common.sh@355 -- # echo 1 00:07:06.727 15:12:57 event -- scripts/common.sh@365 -- # ver1[v]=1 00:07:06.727 15:12:57 event -- scripts/common.sh@366 -- # decimal 2 00:07:06.988 15:12:57 event -- scripts/common.sh@353 -- # local d=2 00:07:06.988 15:12:57 event -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:07:06.988 15:12:57 event -- scripts/common.sh@355 -- # echo 2 00:07:06.988 15:12:57 event -- scripts/common.sh@366 -- # ver2[v]=2 00:07:06.988 15:12:57 event -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:07:06.988 15:12:57 event -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:07:06.988 15:12:57 event -- scripts/common.sh@368 -- # return 0 00:07:06.988 15:12:57 event -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:07:06.988 15:12:57 event -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:07:06.988 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:06.988 --rc genhtml_branch_coverage=1 00:07:06.988 --rc genhtml_function_coverage=1 00:07:06.988 --rc genhtml_legend=1 00:07:06.988 --rc geninfo_all_blocks=1 00:07:06.988 --rc geninfo_unexecuted_blocks=1 00:07:06.988 00:07:06.988 ' 00:07:06.988 15:12:57 event -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:07:06.988 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:06.988 --rc genhtml_branch_coverage=1 00:07:06.988 --rc genhtml_function_coverage=1 00:07:06.988 --rc genhtml_legend=1 00:07:06.988 --rc geninfo_all_blocks=1 00:07:06.988 --rc geninfo_unexecuted_blocks=1 00:07:06.988 00:07:06.988 ' 00:07:06.988 15:12:57 event -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:07:06.988 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:06.988 --rc genhtml_branch_coverage=1 00:07:06.988 --rc genhtml_function_coverage=1 00:07:06.988 --rc genhtml_legend=1 00:07:06.988 --rc geninfo_all_blocks=1 00:07:06.988 --rc geninfo_unexecuted_blocks=1 00:07:06.988 00:07:06.988 ' 00:07:06.988 15:12:57 event -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:07:06.988 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:06.988 --rc genhtml_branch_coverage=1 00:07:06.988 --rc genhtml_function_coverage=1 00:07:06.988 --rc genhtml_legend=1 00:07:06.988 --rc geninfo_all_blocks=1 00:07:06.988 --rc geninfo_unexecuted_blocks=1 00:07:06.988 00:07:06.988 ' 00:07:06.988 15:12:57 event -- event/event.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/bdev/nbd_common.sh 00:07:06.988 15:12:57 event -- bdev/nbd_common.sh@6 -- # set -e 00:07:06.988 15:12:57 event -- event/event.sh@45 -- # run_test event_perf /home/vagrant/spdk_repo/spdk/test/event/event_perf/event_perf -m 0xF -t 1 00:07:06.988 15:12:57 event -- common/autotest_common.sh@1105 -- # '[' 6 -le 1 ']' 00:07:06.988 15:12:57 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:06.988 15:12:57 event -- common/autotest_common.sh@10 -- # set +x 00:07:06.988 ************************************ 00:07:06.988 START TEST event_perf 00:07:06.988 ************************************ 00:07:06.988 15:12:57 event.event_perf -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/event/event_perf/event_perf -m 0xF -t 1 00:07:06.988 Running I/O for 1 seconds...[2024-11-19 15:12:57.126035] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:07:06.988 [2024-11-19 15:12:57.126172] [ DPDK EAL parameters: event_perf --no-shconf -c 0xF --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70006 ] 00:07:06.988 [2024-11-19 15:12:57.282284] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:07:06.988 [2024-11-19 15:12:57.325393] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:07:06.988 [2024-11-19 15:12:57.325616] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:07:06.988 [2024-11-19 15:12:57.325656] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:07.248 [2024-11-19 15:12:57.325796] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:07:08.188 Running I/O for 1 seconds... 00:07:08.188 lcore 0: 93743 00:07:08.188 lcore 1: 93746 00:07:08.188 lcore 2: 93743 00:07:08.188 lcore 3: 93746 00:07:08.188 done. 00:07:08.188 00:07:08.188 real 0m1.328s 00:07:08.188 user 0m4.103s 00:07:08.188 sys 0m0.105s 00:07:08.188 ************************************ 00:07:08.188 END TEST event_perf 00:07:08.188 ************************************ 00:07:08.188 15:12:58 event.event_perf -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:08.188 15:12:58 event.event_perf -- common/autotest_common.sh@10 -- # set +x 00:07:08.188 15:12:58 event -- event/event.sh@46 -- # run_test event_reactor /home/vagrant/spdk_repo/spdk/test/event/reactor/reactor -t 1 00:07:08.188 15:12:58 event -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:07:08.188 15:12:58 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:08.188 15:12:58 event -- common/autotest_common.sh@10 -- # set +x 00:07:08.188 ************************************ 00:07:08.188 START TEST event_reactor 00:07:08.188 ************************************ 00:07:08.188 15:12:58 event.event_reactor -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/event/reactor/reactor -t 1 00:07:08.188 [2024-11-19 15:12:58.523921] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:07:08.188 [2024-11-19 15:12:58.524143] [ DPDK EAL parameters: reactor --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70045 ] 00:07:08.448 [2024-11-19 15:12:58.677978] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:08.448 [2024-11-19 15:12:58.715175] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:09.839 test_start 00:07:09.839 oneshot 00:07:09.839 tick 100 00:07:09.839 tick 100 00:07:09.839 tick 250 00:07:09.839 tick 100 00:07:09.839 tick 100 00:07:09.839 tick 100 00:07:09.839 tick 250 00:07:09.839 tick 500 00:07:09.839 tick 100 00:07:09.839 tick 100 00:07:09.839 tick 250 00:07:09.839 tick 100 00:07:09.839 tick 100 00:07:09.839 test_end 00:07:09.839 00:07:09.839 real 0m1.315s 00:07:09.839 user 0m1.130s 00:07:09.839 sys 0m0.078s 00:07:09.839 ************************************ 00:07:09.839 END TEST event_reactor 00:07:09.839 ************************************ 00:07:09.839 15:12:59 event.event_reactor -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:09.839 15:12:59 event.event_reactor -- common/autotest_common.sh@10 -- # set +x 00:07:09.840 15:12:59 event -- event/event.sh@47 -- # run_test event_reactor_perf /home/vagrant/spdk_repo/spdk/test/event/reactor_perf/reactor_perf -t 1 00:07:09.840 15:12:59 event -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:07:09.840 15:12:59 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:09.840 15:12:59 event -- common/autotest_common.sh@10 -- # set +x 00:07:09.840 ************************************ 00:07:09.840 START TEST event_reactor_perf 00:07:09.840 ************************************ 00:07:09.840 15:12:59 event.event_reactor_perf -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/event/reactor_perf/reactor_perf -t 1 00:07:09.840 [2024-11-19 15:12:59.909427] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:07:09.840 [2024-11-19 15:12:59.909620] [ DPDK EAL parameters: reactor_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70076 ] 00:07:09.840 [2024-11-19 15:13:00.064238] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:09.840 [2024-11-19 15:13:00.105103] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:11.221 test_start 00:07:11.221 test_end 00:07:11.221 Performance: 373089 events per second 00:07:11.221 00:07:11.221 real 0m1.313s 00:07:11.221 user 0m1.124s 00:07:11.221 sys 0m0.081s 00:07:11.221 15:13:01 event.event_reactor_perf -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:11.221 15:13:01 event.event_reactor_perf -- common/autotest_common.sh@10 -- # set +x 00:07:11.221 ************************************ 00:07:11.221 END TEST event_reactor_perf 00:07:11.221 ************************************ 00:07:11.222 15:13:01 event -- event/event.sh@49 -- # uname -s 00:07:11.222 15:13:01 event -- event/event.sh@49 -- # '[' Linux = Linux ']' 00:07:11.222 15:13:01 event -- event/event.sh@50 -- # run_test event_scheduler /home/vagrant/spdk_repo/spdk/test/event/scheduler/scheduler.sh 00:07:11.222 15:13:01 event -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:07:11.222 15:13:01 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:11.222 15:13:01 event -- common/autotest_common.sh@10 -- # set +x 00:07:11.222 ************************************ 00:07:11.222 START TEST event_scheduler 00:07:11.222 ************************************ 00:07:11.222 15:13:01 event.event_scheduler -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/event/scheduler/scheduler.sh 00:07:11.222 * Looking for test storage... 00:07:11.222 * Found test storage at /home/vagrant/spdk_repo/spdk/test/event/scheduler 00:07:11.222 15:13:01 event.event_scheduler -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:07:11.222 15:13:01 event.event_scheduler -- common/autotest_common.sh@1693 -- # lcov --version 00:07:11.222 15:13:01 event.event_scheduler -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:07:11.222 15:13:01 event.event_scheduler -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:07:11.222 15:13:01 event.event_scheduler -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:07:11.222 15:13:01 event.event_scheduler -- scripts/common.sh@333 -- # local ver1 ver1_l 00:07:11.222 15:13:01 event.event_scheduler -- scripts/common.sh@334 -- # local ver2 ver2_l 00:07:11.222 15:13:01 event.event_scheduler -- scripts/common.sh@336 -- # IFS=.-: 00:07:11.222 15:13:01 event.event_scheduler -- scripts/common.sh@336 -- # read -ra ver1 00:07:11.222 15:13:01 event.event_scheduler -- scripts/common.sh@337 -- # IFS=.-: 00:07:11.222 15:13:01 event.event_scheduler -- scripts/common.sh@337 -- # read -ra ver2 00:07:11.222 15:13:01 event.event_scheduler -- scripts/common.sh@338 -- # local 'op=<' 00:07:11.222 15:13:01 event.event_scheduler -- scripts/common.sh@340 -- # ver1_l=2 00:07:11.222 15:13:01 event.event_scheduler -- scripts/common.sh@341 -- # ver2_l=1 00:07:11.222 15:13:01 event.event_scheduler -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:07:11.222 15:13:01 event.event_scheduler -- scripts/common.sh@344 -- # case "$op" in 00:07:11.222 15:13:01 event.event_scheduler -- scripts/common.sh@345 -- # : 1 00:07:11.222 15:13:01 event.event_scheduler -- scripts/common.sh@364 -- # (( v = 0 )) 00:07:11.222 15:13:01 event.event_scheduler -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:07:11.222 15:13:01 event.event_scheduler -- scripts/common.sh@365 -- # decimal 1 00:07:11.222 15:13:01 event.event_scheduler -- scripts/common.sh@353 -- # local d=1 00:07:11.222 15:13:01 event.event_scheduler -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:07:11.222 15:13:01 event.event_scheduler -- scripts/common.sh@355 -- # echo 1 00:07:11.222 15:13:01 event.event_scheduler -- scripts/common.sh@365 -- # ver1[v]=1 00:07:11.222 15:13:01 event.event_scheduler -- scripts/common.sh@366 -- # decimal 2 00:07:11.222 15:13:01 event.event_scheduler -- scripts/common.sh@353 -- # local d=2 00:07:11.222 15:13:01 event.event_scheduler -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:07:11.222 15:13:01 event.event_scheduler -- scripts/common.sh@355 -- # echo 2 00:07:11.222 15:13:01 event.event_scheduler -- scripts/common.sh@366 -- # ver2[v]=2 00:07:11.222 15:13:01 event.event_scheduler -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:07:11.222 15:13:01 event.event_scheduler -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:07:11.222 15:13:01 event.event_scheduler -- scripts/common.sh@368 -- # return 0 00:07:11.222 15:13:01 event.event_scheduler -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:07:11.222 15:13:01 event.event_scheduler -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:07:11.222 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:11.222 --rc genhtml_branch_coverage=1 00:07:11.222 --rc genhtml_function_coverage=1 00:07:11.222 --rc genhtml_legend=1 00:07:11.222 --rc geninfo_all_blocks=1 00:07:11.222 --rc geninfo_unexecuted_blocks=1 00:07:11.222 00:07:11.222 ' 00:07:11.222 15:13:01 event.event_scheduler -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:07:11.222 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:11.222 --rc genhtml_branch_coverage=1 00:07:11.222 --rc genhtml_function_coverage=1 00:07:11.222 --rc genhtml_legend=1 00:07:11.222 --rc geninfo_all_blocks=1 00:07:11.222 --rc geninfo_unexecuted_blocks=1 00:07:11.222 00:07:11.222 ' 00:07:11.222 15:13:01 event.event_scheduler -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:07:11.222 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:11.222 --rc genhtml_branch_coverage=1 00:07:11.222 --rc genhtml_function_coverage=1 00:07:11.222 --rc genhtml_legend=1 00:07:11.222 --rc geninfo_all_blocks=1 00:07:11.222 --rc geninfo_unexecuted_blocks=1 00:07:11.222 00:07:11.222 ' 00:07:11.222 15:13:01 event.event_scheduler -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:07:11.222 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:11.222 --rc genhtml_branch_coverage=1 00:07:11.222 --rc genhtml_function_coverage=1 00:07:11.222 --rc genhtml_legend=1 00:07:11.222 --rc geninfo_all_blocks=1 00:07:11.222 --rc geninfo_unexecuted_blocks=1 00:07:11.222 00:07:11.222 ' 00:07:11.222 15:13:01 event.event_scheduler -- scheduler/scheduler.sh@29 -- # rpc=rpc_cmd 00:07:11.222 15:13:01 event.event_scheduler -- scheduler/scheduler.sh@35 -- # scheduler_pid=70152 00:07:11.222 15:13:01 event.event_scheduler -- scheduler/scheduler.sh@34 -- # /home/vagrant/spdk_repo/spdk/test/event/scheduler/scheduler -m 0xF -p 0x2 --wait-for-rpc -f 00:07:11.222 15:13:01 event.event_scheduler -- scheduler/scheduler.sh@36 -- # trap 'killprocess $scheduler_pid; exit 1' SIGINT SIGTERM EXIT 00:07:11.222 15:13:01 event.event_scheduler -- scheduler/scheduler.sh@37 -- # waitforlisten 70152 00:07:11.222 15:13:01 event.event_scheduler -- common/autotest_common.sh@835 -- # '[' -z 70152 ']' 00:07:11.222 15:13:01 event.event_scheduler -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:11.222 15:13:01 event.event_scheduler -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:11.222 15:13:01 event.event_scheduler -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:11.222 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:11.222 15:13:01 event.event_scheduler -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:11.222 15:13:01 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:07:11.483 [2024-11-19 15:13:01.567450] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:07:11.483 [2024-11-19 15:13:01.567693] [ DPDK EAL parameters: scheduler --no-shconf -c 0xF --main-lcore=2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70152 ] 00:07:11.483 [2024-11-19 15:13:01.706738] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:07:11.483 [2024-11-19 15:13:01.750154] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:11.483 [2024-11-19 15:13:01.750337] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:07:11.483 [2024-11-19 15:13:01.750420] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:07:11.483 [2024-11-19 15:13:01.750537] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:07:12.422 15:13:02 event.event_scheduler -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:12.422 15:13:02 event.event_scheduler -- common/autotest_common.sh@868 -- # return 0 00:07:12.422 15:13:02 event.event_scheduler -- scheduler/scheduler.sh@39 -- # rpc_cmd framework_set_scheduler dynamic 00:07:12.422 15:13:02 event.event_scheduler -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:12.422 15:13:02 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:07:12.422 POWER: failed to open /sys/devices/system/cpu/cpu%u/cpufreq/scaling_governor 00:07:12.422 POWER: Cannot set governor of lcore 0 to userspace 00:07:12.422 POWER: failed to open /sys/devices/system/cpu/cpu%u/cpufreq/scaling_governor 00:07:12.422 POWER: Cannot set governor of lcore 0 to performance 00:07:12.422 POWER: failed to open /sys/devices/system/cpu/cpu%u/cpufreq/scaling_governor 00:07:12.422 POWER: Cannot set governor of lcore 0 to userspace 00:07:12.422 GUEST_CHANNEL: Unable to to connect to '/dev/virtio-ports/virtio.serial.port.poweragent.0' with error No such file or directory 00:07:12.422 POWER: Unable to set Power Management Environment for lcore 0 00:07:12.422 [2024-11-19 15:13:02.439563] dpdk_governor.c: 130:_init_core: *ERROR*: Failed to initialize on core0 00:07:12.422 [2024-11-19 15:13:02.439614] dpdk_governor.c: 191:_init: *ERROR*: Failed to initialize on core0 00:07:12.422 [2024-11-19 15:13:02.439676] scheduler_dynamic.c: 280:init: *NOTICE*: Unable to initialize dpdk governor 00:07:12.422 [2024-11-19 15:13:02.439752] scheduler_dynamic.c: 427:set_opts: *NOTICE*: Setting scheduler load limit to 20 00:07:12.422 [2024-11-19 15:13:02.439800] scheduler_dynamic.c: 429:set_opts: *NOTICE*: Setting scheduler core limit to 80 00:07:12.422 [2024-11-19 15:13:02.439855] scheduler_dynamic.c: 431:set_opts: *NOTICE*: Setting scheduler core busy to 95 00:07:12.422 15:13:02 event.event_scheduler -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:12.422 15:13:02 event.event_scheduler -- scheduler/scheduler.sh@40 -- # rpc_cmd framework_start_init 00:07:12.422 15:13:02 event.event_scheduler -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:12.422 15:13:02 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:07:12.422 [2024-11-19 15:13:02.569331] scheduler.c: 382:test_start: *NOTICE*: Scheduler test application started. 00:07:12.422 15:13:02 event.event_scheduler -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:12.422 15:13:02 event.event_scheduler -- scheduler/scheduler.sh@43 -- # run_test scheduler_create_thread scheduler_create_thread 00:07:12.422 15:13:02 event.event_scheduler -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:07:12.422 15:13:02 event.event_scheduler -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:12.422 15:13:02 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:07:12.422 ************************************ 00:07:12.422 START TEST scheduler_create_thread 00:07:12.422 ************************************ 00:07:12.422 15:13:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@1129 -- # scheduler_create_thread 00:07:12.422 15:13:02 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@12 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x1 -a 100 00:07:12.422 15:13:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:12.423 15:13:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:07:12.423 2 00:07:12.423 15:13:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:12.423 15:13:02 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@13 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x2 -a 100 00:07:12.423 15:13:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:12.423 15:13:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:07:12.423 3 00:07:12.423 15:13:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:12.423 15:13:02 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@14 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x4 -a 100 00:07:12.423 15:13:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:12.423 15:13:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:07:12.423 4 00:07:12.423 15:13:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:12.423 15:13:02 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@15 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x8 -a 100 00:07:12.423 15:13:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:12.423 15:13:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:07:12.423 5 00:07:12.423 15:13:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:12.423 15:13:02 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@16 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x1 -a 0 00:07:12.423 15:13:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:12.423 15:13:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:07:12.423 6 00:07:12.423 15:13:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:12.423 15:13:02 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@17 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x2 -a 0 00:07:12.423 15:13:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:12.423 15:13:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:07:12.423 7 00:07:12.423 15:13:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:12.423 15:13:02 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@18 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x4 -a 0 00:07:12.423 15:13:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:12.423 15:13:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:07:12.423 8 00:07:12.423 15:13:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:12.423 15:13:02 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@19 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x8 -a 0 00:07:12.423 15:13:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:12.423 15:13:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:07:12.423 9 00:07:12.423 15:13:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:12.423 15:13:02 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@21 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n one_third_active -a 30 00:07:12.423 15:13:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:12.423 15:13:02 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:07:12.993 10 00:07:12.993 15:13:03 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:12.993 15:13:03 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@22 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n half_active -a 0 00:07:12.993 15:13:03 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:12.993 15:13:03 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:07:14.374 15:13:04 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:14.374 15:13:04 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@22 -- # thread_id=11 00:07:14.374 15:13:04 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@23 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_set_active 11 50 00:07:14.374 15:13:04 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:14.374 15:13:04 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:07:14.943 15:13:05 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:14.943 15:13:05 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@25 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n deleted -a 100 00:07:14.943 15:13:05 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:14.943 15:13:05 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:07:15.882 15:13:06 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:15.882 15:13:06 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@25 -- # thread_id=12 00:07:15.882 15:13:06 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@26 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_delete 12 00:07:15.882 15:13:06 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:15.882 15:13:06 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:07:16.821 ************************************ 00:07:16.821 END TEST scheduler_create_thread 00:07:16.821 ************************************ 00:07:16.821 15:13:06 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:16.821 00:07:16.821 real 0m4.212s 00:07:16.821 user 0m0.032s 00:07:16.821 sys 0m0.006s 00:07:16.821 15:13:06 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:16.821 15:13:06 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:07:16.821 15:13:06 event.event_scheduler -- scheduler/scheduler.sh@45 -- # trap - SIGINT SIGTERM EXIT 00:07:16.821 15:13:06 event.event_scheduler -- scheduler/scheduler.sh@46 -- # killprocess 70152 00:07:16.821 15:13:06 event.event_scheduler -- common/autotest_common.sh@954 -- # '[' -z 70152 ']' 00:07:16.821 15:13:06 event.event_scheduler -- common/autotest_common.sh@958 -- # kill -0 70152 00:07:16.821 15:13:06 event.event_scheduler -- common/autotest_common.sh@959 -- # uname 00:07:16.821 15:13:06 event.event_scheduler -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:16.821 15:13:06 event.event_scheduler -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70152 00:07:16.821 killing process with pid 70152 00:07:16.821 15:13:06 event.event_scheduler -- common/autotest_common.sh@960 -- # process_name=reactor_2 00:07:16.821 15:13:06 event.event_scheduler -- common/autotest_common.sh@964 -- # '[' reactor_2 = sudo ']' 00:07:16.821 15:13:06 event.event_scheduler -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70152' 00:07:16.821 15:13:06 event.event_scheduler -- common/autotest_common.sh@973 -- # kill 70152 00:07:16.821 15:13:06 event.event_scheduler -- common/autotest_common.sh@978 -- # wait 70152 00:07:17.080 [2024-11-19 15:13:07.175216] scheduler.c: 360:test_shutdown: *NOTICE*: Scheduler test application stopped. 00:07:17.340 00:07:17.340 real 0m6.298s 00:07:17.340 user 0m14.176s 00:07:17.340 sys 0m0.559s 00:07:17.340 15:13:07 event.event_scheduler -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:17.340 15:13:07 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:07:17.340 ************************************ 00:07:17.340 END TEST event_scheduler 00:07:17.340 ************************************ 00:07:17.340 15:13:07 event -- event/event.sh@51 -- # modprobe -n nbd 00:07:17.340 15:13:07 event -- event/event.sh@52 -- # run_test app_repeat app_repeat_test 00:07:17.340 15:13:07 event -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:07:17.340 15:13:07 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:17.340 15:13:07 event -- common/autotest_common.sh@10 -- # set +x 00:07:17.340 ************************************ 00:07:17.340 START TEST app_repeat 00:07:17.340 ************************************ 00:07:17.340 15:13:07 event.app_repeat -- common/autotest_common.sh@1129 -- # app_repeat_test 00:07:17.340 15:13:07 event.app_repeat -- event/event.sh@12 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:17.340 15:13:07 event.app_repeat -- event/event.sh@13 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:17.340 15:13:07 event.app_repeat -- event/event.sh@13 -- # local nbd_list 00:07:17.340 15:13:07 event.app_repeat -- event/event.sh@14 -- # bdev_list=('Malloc0' 'Malloc1') 00:07:17.340 15:13:07 event.app_repeat -- event/event.sh@14 -- # local bdev_list 00:07:17.340 15:13:07 event.app_repeat -- event/event.sh@15 -- # local repeat_times=4 00:07:17.340 15:13:07 event.app_repeat -- event/event.sh@17 -- # modprobe nbd 00:07:17.340 15:13:07 event.app_repeat -- event/event.sh@19 -- # repeat_pid=70265 00:07:17.340 Process app_repeat pid: 70265 00:07:17.340 15:13:07 event.app_repeat -- event/event.sh@18 -- # /home/vagrant/spdk_repo/spdk/test/event/app_repeat/app_repeat -r /var/tmp/spdk-nbd.sock -m 0x3 -t 4 00:07:17.340 15:13:07 event.app_repeat -- event/event.sh@20 -- # trap 'killprocess $repeat_pid; exit 1' SIGINT SIGTERM EXIT 00:07:17.340 15:13:07 event.app_repeat -- event/event.sh@21 -- # echo 'Process app_repeat pid: 70265' 00:07:17.340 spdk_app_start Round 0 00:07:17.340 15:13:07 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:07:17.340 15:13:07 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 0' 00:07:17.340 15:13:07 event.app_repeat -- event/event.sh@25 -- # waitforlisten 70265 /var/tmp/spdk-nbd.sock 00:07:17.340 15:13:07 event.app_repeat -- common/autotest_common.sh@835 -- # '[' -z 70265 ']' 00:07:17.340 15:13:07 event.app_repeat -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:07:17.340 15:13:07 event.app_repeat -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:17.340 15:13:07 event.app_repeat -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:07:17.340 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:07:17.340 15:13:07 event.app_repeat -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:17.340 15:13:07 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:07:17.599 [2024-11-19 15:13:07.686099] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:07:17.599 [2024-11-19 15:13:07.686325] [ DPDK EAL parameters: app_repeat --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70265 ] 00:07:17.599 [2024-11-19 15:13:07.841648] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:07:17.599 [2024-11-19 15:13:07.881519] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:17.599 [2024-11-19 15:13:07.881627] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:07:18.538 15:13:08 event.app_repeat -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:18.538 15:13:08 event.app_repeat -- common/autotest_common.sh@868 -- # return 0 00:07:18.538 15:13:08 event.app_repeat -- event/event.sh@27 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:07:18.538 Malloc0 00:07:18.538 15:13:08 event.app_repeat -- event/event.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:07:18.798 Malloc1 00:07:18.798 15:13:09 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:07:18.798 15:13:09 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:18.798 15:13:09 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:07:18.798 15:13:09 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:07:18.798 15:13:09 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:18.798 15:13:09 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:07:18.798 15:13:09 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:07:18.798 15:13:09 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:18.798 15:13:09 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:07:18.798 15:13:09 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:07:18.798 15:13:09 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:18.798 15:13:09 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:07:18.798 15:13:09 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:07:18.798 15:13:09 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:07:18.798 15:13:09 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:07:18.798 15:13:09 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:07:19.056 /dev/nbd0 00:07:19.056 15:13:09 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:07:19.056 15:13:09 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:07:19.056 15:13:09 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:07:19.056 15:13:09 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:07:19.056 15:13:09 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:07:19.056 15:13:09 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:07:19.056 15:13:09 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:07:19.056 15:13:09 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:07:19.056 15:13:09 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:07:19.056 15:13:09 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:07:19.056 15:13:09 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:07:19.056 1+0 records in 00:07:19.056 1+0 records out 00:07:19.056 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000348211 s, 11.8 MB/s 00:07:19.056 15:13:09 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:07:19.056 15:13:09 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:07:19.056 15:13:09 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:07:19.056 15:13:09 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:07:19.056 15:13:09 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:07:19.056 15:13:09 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:07:19.056 15:13:09 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:07:19.056 15:13:09 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:07:19.352 /dev/nbd1 00:07:19.352 15:13:09 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:07:19.352 15:13:09 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:07:19.352 15:13:09 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:07:19.352 15:13:09 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:07:19.352 15:13:09 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:07:19.352 15:13:09 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:07:19.352 15:13:09 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:07:19.352 15:13:09 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:07:19.352 15:13:09 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:07:19.352 15:13:09 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:07:19.352 15:13:09 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:07:19.352 1+0 records in 00:07:19.352 1+0 records out 00:07:19.352 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000456423 s, 9.0 MB/s 00:07:19.352 15:13:09 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:07:19.352 15:13:09 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:07:19.352 15:13:09 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:07:19.352 15:13:09 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:07:19.352 15:13:09 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:07:19.352 15:13:09 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:07:19.352 15:13:09 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:07:19.352 15:13:09 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:07:19.352 15:13:09 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:19.352 15:13:09 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:07:19.632 15:13:09 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:07:19.632 { 00:07:19.632 "nbd_device": "/dev/nbd0", 00:07:19.632 "bdev_name": "Malloc0" 00:07:19.632 }, 00:07:19.632 { 00:07:19.632 "nbd_device": "/dev/nbd1", 00:07:19.632 "bdev_name": "Malloc1" 00:07:19.632 } 00:07:19.632 ]' 00:07:19.632 15:13:09 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:07:19.632 { 00:07:19.632 "nbd_device": "/dev/nbd0", 00:07:19.632 "bdev_name": "Malloc0" 00:07:19.632 }, 00:07:19.632 { 00:07:19.632 "nbd_device": "/dev/nbd1", 00:07:19.632 "bdev_name": "Malloc1" 00:07:19.632 } 00:07:19.632 ]' 00:07:19.632 15:13:09 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:07:19.632 15:13:09 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:07:19.632 /dev/nbd1' 00:07:19.632 15:13:09 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:07:19.632 15:13:09 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:07:19.632 /dev/nbd1' 00:07:19.632 15:13:09 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:07:19.632 15:13:09 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:07:19.632 15:13:09 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:07:19.632 15:13:09 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:07:19.632 15:13:09 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:07:19.632 15:13:09 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:19.632 15:13:09 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:07:19.632 15:13:09 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:07:19.633 15:13:09 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:07:19.633 15:13:09 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:07:19.633 15:13:09 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest bs=4096 count=256 00:07:19.633 256+0 records in 00:07:19.633 256+0 records out 00:07:19.633 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0122526 s, 85.6 MB/s 00:07:19.633 15:13:09 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:07:19.633 15:13:09 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:07:19.633 256+0 records in 00:07:19.633 256+0 records out 00:07:19.633 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0243026 s, 43.1 MB/s 00:07:19.633 15:13:09 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:07:19.633 15:13:09 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:07:19.633 256+0 records in 00:07:19.633 256+0 records out 00:07:19.633 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0244553 s, 42.9 MB/s 00:07:19.633 15:13:09 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:07:19.633 15:13:09 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:19.633 15:13:09 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:07:19.633 15:13:09 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:07:19.633 15:13:09 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:07:19.633 15:13:09 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:07:19.633 15:13:09 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:07:19.633 15:13:09 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:07:19.633 15:13:09 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd0 00:07:19.633 15:13:09 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:07:19.633 15:13:09 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd1 00:07:19.633 15:13:09 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:07:19.633 15:13:09 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:07:19.633 15:13:09 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:19.633 15:13:09 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:19.633 15:13:09 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:07:19.633 15:13:09 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:07:19.633 15:13:09 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:07:19.633 15:13:09 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:07:19.892 15:13:10 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:07:19.892 15:13:10 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:07:19.892 15:13:10 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:07:19.892 15:13:10 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:07:19.892 15:13:10 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:07:19.892 15:13:10 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:07:19.892 15:13:10 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:07:19.892 15:13:10 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:07:19.892 15:13:10 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:07:19.892 15:13:10 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:07:20.152 15:13:10 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:07:20.152 15:13:10 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:07:20.152 15:13:10 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:07:20.152 15:13:10 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:07:20.152 15:13:10 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:07:20.152 15:13:10 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:07:20.152 15:13:10 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:07:20.152 15:13:10 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:07:20.152 15:13:10 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:07:20.152 15:13:10 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:20.152 15:13:10 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:07:20.412 15:13:10 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:07:20.412 15:13:10 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:07:20.412 15:13:10 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:07:20.412 15:13:10 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:07:20.412 15:13:10 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:07:20.412 15:13:10 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:07:20.412 15:13:10 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:07:20.412 15:13:10 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:07:20.412 15:13:10 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:07:20.412 15:13:10 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:07:20.412 15:13:10 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:07:20.412 15:13:10 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:07:20.412 15:13:10 event.app_repeat -- event/event.sh@34 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:07:20.672 15:13:10 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:07:20.932 [2024-11-19 15:13:11.066001] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:07:20.932 [2024-11-19 15:13:11.100330] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:20.932 [2024-11-19 15:13:11.100332] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:07:20.932 [2024-11-19 15:13:11.176193] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:07:20.932 [2024-11-19 15:13:11.176269] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:07:24.228 spdk_app_start Round 1 00:07:24.228 15:13:13 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:07:24.228 15:13:13 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 1' 00:07:24.228 15:13:13 event.app_repeat -- event/event.sh@25 -- # waitforlisten 70265 /var/tmp/spdk-nbd.sock 00:07:24.228 15:13:13 event.app_repeat -- common/autotest_common.sh@835 -- # '[' -z 70265 ']' 00:07:24.228 15:13:13 event.app_repeat -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:07:24.228 15:13:13 event.app_repeat -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:24.228 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:07:24.228 15:13:13 event.app_repeat -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:07:24.228 15:13:13 event.app_repeat -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:24.228 15:13:13 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:07:24.228 15:13:14 event.app_repeat -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:24.228 15:13:14 event.app_repeat -- common/autotest_common.sh@868 -- # return 0 00:07:24.228 15:13:14 event.app_repeat -- event/event.sh@27 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:07:24.228 Malloc0 00:07:24.228 15:13:14 event.app_repeat -- event/event.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:07:24.228 Malloc1 00:07:24.228 15:13:14 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:07:24.228 15:13:14 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:24.228 15:13:14 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:07:24.228 15:13:14 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:07:24.228 15:13:14 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:24.228 15:13:14 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:07:24.228 15:13:14 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:07:24.228 15:13:14 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:24.228 15:13:14 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:07:24.228 15:13:14 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:07:24.228 15:13:14 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:24.228 15:13:14 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:07:24.228 15:13:14 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:07:24.228 15:13:14 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:07:24.228 15:13:14 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:07:24.228 15:13:14 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:07:24.488 /dev/nbd0 00:07:24.489 15:13:14 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:07:24.489 15:13:14 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:07:24.489 15:13:14 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:07:24.489 15:13:14 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:07:24.489 15:13:14 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:07:24.489 15:13:14 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:07:24.489 15:13:14 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:07:24.489 15:13:14 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:07:24.489 15:13:14 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:07:24.489 15:13:14 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:07:24.489 15:13:14 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:07:24.489 1+0 records in 00:07:24.489 1+0 records out 00:07:24.489 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000377615 s, 10.8 MB/s 00:07:24.489 15:13:14 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:07:24.489 15:13:14 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:07:24.489 15:13:14 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:07:24.489 15:13:14 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:07:24.489 15:13:14 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:07:24.489 15:13:14 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:07:24.489 15:13:14 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:07:24.489 15:13:14 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:07:24.749 /dev/nbd1 00:07:24.749 15:13:14 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:07:24.749 15:13:14 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:07:24.749 15:13:14 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:07:24.749 15:13:14 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:07:24.749 15:13:14 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:07:24.749 15:13:14 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:07:24.749 15:13:14 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:07:24.749 15:13:14 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:07:24.749 15:13:14 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:07:24.749 15:13:14 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:07:24.749 15:13:14 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:07:24.749 1+0 records in 00:07:24.749 1+0 records out 00:07:24.749 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000201659 s, 20.3 MB/s 00:07:24.749 15:13:14 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:07:24.749 15:13:14 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:07:24.749 15:13:14 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:07:24.749 15:13:14 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:07:24.749 15:13:14 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:07:24.749 15:13:14 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:07:24.749 15:13:14 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:07:24.749 15:13:14 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:07:24.749 15:13:14 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:24.749 15:13:14 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:07:25.010 15:13:15 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:07:25.010 { 00:07:25.010 "nbd_device": "/dev/nbd0", 00:07:25.010 "bdev_name": "Malloc0" 00:07:25.010 }, 00:07:25.010 { 00:07:25.010 "nbd_device": "/dev/nbd1", 00:07:25.010 "bdev_name": "Malloc1" 00:07:25.010 } 00:07:25.010 ]' 00:07:25.010 15:13:15 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:07:25.010 { 00:07:25.010 "nbd_device": "/dev/nbd0", 00:07:25.010 "bdev_name": "Malloc0" 00:07:25.010 }, 00:07:25.010 { 00:07:25.010 "nbd_device": "/dev/nbd1", 00:07:25.010 "bdev_name": "Malloc1" 00:07:25.010 } 00:07:25.010 ]' 00:07:25.010 15:13:15 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:07:25.010 15:13:15 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:07:25.010 /dev/nbd1' 00:07:25.010 15:13:15 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:07:25.010 15:13:15 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:07:25.010 /dev/nbd1' 00:07:25.010 15:13:15 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:07:25.010 15:13:15 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:07:25.010 15:13:15 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:07:25.010 15:13:15 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:07:25.010 15:13:15 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:07:25.010 15:13:15 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:25.010 15:13:15 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:07:25.010 15:13:15 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:07:25.010 15:13:15 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:07:25.010 15:13:15 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:07:25.010 15:13:15 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest bs=4096 count=256 00:07:25.010 256+0 records in 00:07:25.010 256+0 records out 00:07:25.010 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0122293 s, 85.7 MB/s 00:07:25.010 15:13:15 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:07:25.010 15:13:15 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:07:25.010 256+0 records in 00:07:25.010 256+0 records out 00:07:25.010 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0183016 s, 57.3 MB/s 00:07:25.010 15:13:15 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:07:25.010 15:13:15 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:07:25.010 256+0 records in 00:07:25.010 256+0 records out 00:07:25.010 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0213253 s, 49.2 MB/s 00:07:25.010 15:13:15 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:07:25.010 15:13:15 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:25.010 15:13:15 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:07:25.010 15:13:15 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:07:25.010 15:13:15 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:07:25.010 15:13:15 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:07:25.010 15:13:15 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:07:25.010 15:13:15 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:07:25.010 15:13:15 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd0 00:07:25.010 15:13:15 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:07:25.010 15:13:15 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd1 00:07:25.010 15:13:15 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:07:25.010 15:13:15 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:07:25.010 15:13:15 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:25.010 15:13:15 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:25.010 15:13:15 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:07:25.010 15:13:15 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:07:25.010 15:13:15 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:07:25.010 15:13:15 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:07:25.271 15:13:15 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:07:25.271 15:13:15 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:07:25.271 15:13:15 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:07:25.271 15:13:15 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:07:25.271 15:13:15 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:07:25.271 15:13:15 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:07:25.271 15:13:15 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:07:25.271 15:13:15 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:07:25.271 15:13:15 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:07:25.271 15:13:15 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:07:25.531 15:13:15 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:07:25.531 15:13:15 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:07:25.531 15:13:15 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:07:25.531 15:13:15 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:07:25.531 15:13:15 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:07:25.531 15:13:15 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:07:25.531 15:13:15 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:07:25.531 15:13:15 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:07:25.531 15:13:15 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:07:25.531 15:13:15 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:25.531 15:13:15 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:07:25.791 15:13:15 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:07:25.791 15:13:15 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:07:25.791 15:13:15 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:07:25.791 15:13:16 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:07:25.791 15:13:16 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:07:25.791 15:13:16 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:07:25.791 15:13:16 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:07:25.791 15:13:16 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:07:25.791 15:13:16 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:07:25.791 15:13:16 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:07:25.791 15:13:16 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:07:25.791 15:13:16 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:07:25.791 15:13:16 event.app_repeat -- event/event.sh@34 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:07:26.051 15:13:16 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:07:26.312 [2024-11-19 15:13:16.563208] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:07:26.312 [2024-11-19 15:13:16.598052] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:26.312 [2024-11-19 15:13:16.598091] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:07:26.572 [2024-11-19 15:13:16.673761] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:07:26.572 [2024-11-19 15:13:16.673832] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:07:29.113 spdk_app_start Round 2 00:07:29.113 15:13:19 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:07:29.113 15:13:19 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 2' 00:07:29.113 15:13:19 event.app_repeat -- event/event.sh@25 -- # waitforlisten 70265 /var/tmp/spdk-nbd.sock 00:07:29.113 15:13:19 event.app_repeat -- common/autotest_common.sh@835 -- # '[' -z 70265 ']' 00:07:29.113 15:13:19 event.app_repeat -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:07:29.113 15:13:19 event.app_repeat -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:29.113 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:07:29.113 15:13:19 event.app_repeat -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:07:29.113 15:13:19 event.app_repeat -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:29.114 15:13:19 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:07:29.373 15:13:19 event.app_repeat -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:29.373 15:13:19 event.app_repeat -- common/autotest_common.sh@868 -- # return 0 00:07:29.373 15:13:19 event.app_repeat -- event/event.sh@27 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:07:29.633 Malloc0 00:07:29.633 15:13:19 event.app_repeat -- event/event.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:07:29.633 Malloc1 00:07:29.893 15:13:19 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:07:29.893 15:13:19 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:29.893 15:13:19 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:07:29.893 15:13:19 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:07:29.893 15:13:19 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:29.893 15:13:19 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:07:29.893 15:13:19 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:07:29.893 15:13:19 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:29.893 15:13:19 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:07:29.893 15:13:19 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:07:29.893 15:13:19 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:29.893 15:13:19 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:07:29.893 15:13:19 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:07:29.893 15:13:19 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:07:29.893 15:13:19 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:07:29.893 15:13:19 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:07:29.893 /dev/nbd0 00:07:29.893 15:13:20 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:07:29.893 15:13:20 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:07:29.893 15:13:20 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:07:29.893 15:13:20 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:07:29.893 15:13:20 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:07:29.893 15:13:20 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:07:29.893 15:13:20 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:07:29.893 15:13:20 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:07:29.893 15:13:20 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:07:29.893 15:13:20 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:07:29.893 15:13:20 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:07:29.893 1+0 records in 00:07:29.893 1+0 records out 00:07:29.893 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000467502 s, 8.8 MB/s 00:07:29.893 15:13:20 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:07:29.893 15:13:20 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:07:29.893 15:13:20 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:07:29.893 15:13:20 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:07:29.893 15:13:20 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:07:29.893 15:13:20 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:07:29.893 15:13:20 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:07:29.893 15:13:20 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:07:30.154 /dev/nbd1 00:07:30.154 15:13:20 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:07:30.154 15:13:20 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:07:30.154 15:13:20 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:07:30.154 15:13:20 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:07:30.154 15:13:20 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:07:30.154 15:13:20 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:07:30.154 15:13:20 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:07:30.154 15:13:20 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:07:30.154 15:13:20 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:07:30.154 15:13:20 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:07:30.154 15:13:20 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:07:30.154 1+0 records in 00:07:30.154 1+0 records out 00:07:30.154 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000386155 s, 10.6 MB/s 00:07:30.154 15:13:20 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:07:30.154 15:13:20 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:07:30.154 15:13:20 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:07:30.154 15:13:20 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:07:30.154 15:13:20 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:07:30.154 15:13:20 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:07:30.154 15:13:20 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:07:30.154 15:13:20 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:07:30.154 15:13:20 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:30.154 15:13:20 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:07:30.414 15:13:20 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:07:30.414 { 00:07:30.414 "nbd_device": "/dev/nbd0", 00:07:30.414 "bdev_name": "Malloc0" 00:07:30.414 }, 00:07:30.414 { 00:07:30.414 "nbd_device": "/dev/nbd1", 00:07:30.414 "bdev_name": "Malloc1" 00:07:30.414 } 00:07:30.414 ]' 00:07:30.414 15:13:20 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:07:30.414 { 00:07:30.414 "nbd_device": "/dev/nbd0", 00:07:30.414 "bdev_name": "Malloc0" 00:07:30.414 }, 00:07:30.414 { 00:07:30.414 "nbd_device": "/dev/nbd1", 00:07:30.414 "bdev_name": "Malloc1" 00:07:30.414 } 00:07:30.414 ]' 00:07:30.414 15:13:20 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:07:30.414 15:13:20 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:07:30.414 /dev/nbd1' 00:07:30.414 15:13:20 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:07:30.414 /dev/nbd1' 00:07:30.414 15:13:20 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:07:30.414 15:13:20 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:07:30.414 15:13:20 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:07:30.414 15:13:20 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:07:30.414 15:13:20 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:07:30.414 15:13:20 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:07:30.414 15:13:20 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:30.414 15:13:20 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:07:30.414 15:13:20 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:07:30.414 15:13:20 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:07:30.414 15:13:20 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:07:30.414 15:13:20 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest bs=4096 count=256 00:07:30.681 256+0 records in 00:07:30.681 256+0 records out 00:07:30.681 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0128173 s, 81.8 MB/s 00:07:30.681 15:13:20 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:07:30.681 15:13:20 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:07:30.681 256+0 records in 00:07:30.681 256+0 records out 00:07:30.681 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0198754 s, 52.8 MB/s 00:07:30.681 15:13:20 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:07:30.681 15:13:20 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:07:30.681 256+0 records in 00:07:30.681 256+0 records out 00:07:30.681 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0242946 s, 43.2 MB/s 00:07:30.681 15:13:20 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:07:30.681 15:13:20 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:30.681 15:13:20 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:07:30.681 15:13:20 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:07:30.681 15:13:20 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:07:30.681 15:13:20 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:07:30.681 15:13:20 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:07:30.681 15:13:20 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:07:30.681 15:13:20 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd0 00:07:30.681 15:13:20 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:07:30.681 15:13:20 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd1 00:07:30.681 15:13:20 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:07:30.681 15:13:20 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:07:30.681 15:13:20 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:30.681 15:13:20 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:07:30.681 15:13:20 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:07:30.681 15:13:20 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:07:30.681 15:13:20 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:07:30.681 15:13:20 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:07:30.968 15:13:21 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:07:30.968 15:13:21 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:07:30.968 15:13:21 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:07:30.968 15:13:21 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:07:30.968 15:13:21 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:07:30.968 15:13:21 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:07:30.968 15:13:21 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:07:30.968 15:13:21 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:07:30.968 15:13:21 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:07:30.968 15:13:21 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:07:30.968 15:13:21 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:07:30.968 15:13:21 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:07:30.968 15:13:21 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:07:30.968 15:13:21 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:07:30.968 15:13:21 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:07:30.968 15:13:21 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:07:30.968 15:13:21 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:07:30.968 15:13:21 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:07:30.968 15:13:21 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:07:30.968 15:13:21 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:07:30.968 15:13:21 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:07:31.239 15:13:21 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:07:31.239 15:13:21 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:07:31.239 15:13:21 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:07:31.239 15:13:21 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:07:31.239 15:13:21 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:07:31.239 15:13:21 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:07:31.239 15:13:21 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:07:31.239 15:13:21 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:07:31.239 15:13:21 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:07:31.239 15:13:21 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:07:31.239 15:13:21 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:07:31.239 15:13:21 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:07:31.239 15:13:21 event.app_repeat -- event/event.sh@34 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:07:31.499 15:13:21 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:07:31.759 [2024-11-19 15:13:22.025191] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:07:31.759 [2024-11-19 15:13:22.060146] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:31.759 [2024-11-19 15:13:22.060150] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:07:32.024 [2024-11-19 15:13:22.138664] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:07:32.024 [2024-11-19 15:13:22.138737] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:07:34.565 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:07:34.565 15:13:24 event.app_repeat -- event/event.sh@38 -- # waitforlisten 70265 /var/tmp/spdk-nbd.sock 00:07:34.565 15:13:24 event.app_repeat -- common/autotest_common.sh@835 -- # '[' -z 70265 ']' 00:07:34.565 15:13:24 event.app_repeat -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:07:34.565 15:13:24 event.app_repeat -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:34.565 15:13:24 event.app_repeat -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:07:34.565 15:13:24 event.app_repeat -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:34.565 15:13:24 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:07:34.825 15:13:25 event.app_repeat -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:34.825 15:13:25 event.app_repeat -- common/autotest_common.sh@868 -- # return 0 00:07:34.825 15:13:25 event.app_repeat -- event/event.sh@39 -- # killprocess 70265 00:07:34.825 15:13:25 event.app_repeat -- common/autotest_common.sh@954 -- # '[' -z 70265 ']' 00:07:34.825 15:13:25 event.app_repeat -- common/autotest_common.sh@958 -- # kill -0 70265 00:07:34.825 15:13:25 event.app_repeat -- common/autotest_common.sh@959 -- # uname 00:07:34.825 15:13:25 event.app_repeat -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:34.825 15:13:25 event.app_repeat -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70265 00:07:34.825 15:13:25 event.app_repeat -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:34.825 15:13:25 event.app_repeat -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:34.825 15:13:25 event.app_repeat -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70265' 00:07:34.825 killing process with pid 70265 00:07:34.825 15:13:25 event.app_repeat -- common/autotest_common.sh@973 -- # kill 70265 00:07:34.825 15:13:25 event.app_repeat -- common/autotest_common.sh@978 -- # wait 70265 00:07:35.085 spdk_app_start is called in Round 0. 00:07:35.085 Shutdown signal received, stop current app iteration 00:07:35.085 Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 reinitialization... 00:07:35.085 spdk_app_start is called in Round 1. 00:07:35.085 Shutdown signal received, stop current app iteration 00:07:35.085 Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 reinitialization... 00:07:35.085 spdk_app_start is called in Round 2. 00:07:35.085 Shutdown signal received, stop current app iteration 00:07:35.085 Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 reinitialization... 00:07:35.085 spdk_app_start is called in Round 3. 00:07:35.085 Shutdown signal received, stop current app iteration 00:07:35.085 15:13:25 event.app_repeat -- event/event.sh@40 -- # trap - SIGINT SIGTERM EXIT 00:07:35.085 15:13:25 event.app_repeat -- event/event.sh@42 -- # return 0 00:07:35.085 00:07:35.085 real 0m17.691s 00:07:35.085 user 0m38.697s 00:07:35.085 sys 0m2.950s 00:07:35.085 ************************************ 00:07:35.085 END TEST app_repeat 00:07:35.085 ************************************ 00:07:35.085 15:13:25 event.app_repeat -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:35.085 15:13:25 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:07:35.085 15:13:25 event -- event/event.sh@54 -- # (( SPDK_TEST_CRYPTO == 0 )) 00:07:35.085 15:13:25 event -- event/event.sh@55 -- # run_test cpu_locks /home/vagrant/spdk_repo/spdk/test/event/cpu_locks.sh 00:07:35.085 15:13:25 event -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:07:35.085 15:13:25 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:35.085 15:13:25 event -- common/autotest_common.sh@10 -- # set +x 00:07:35.085 ************************************ 00:07:35.085 START TEST cpu_locks 00:07:35.085 ************************************ 00:07:35.085 15:13:25 event.cpu_locks -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/event/cpu_locks.sh 00:07:35.345 * Looking for test storage... 00:07:35.346 * Found test storage at /home/vagrant/spdk_repo/spdk/test/event 00:07:35.346 15:13:25 event.cpu_locks -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:07:35.346 15:13:25 event.cpu_locks -- common/autotest_common.sh@1693 -- # lcov --version 00:07:35.346 15:13:25 event.cpu_locks -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:07:35.346 15:13:25 event.cpu_locks -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:07:35.346 15:13:25 event.cpu_locks -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:07:35.346 15:13:25 event.cpu_locks -- scripts/common.sh@333 -- # local ver1 ver1_l 00:07:35.346 15:13:25 event.cpu_locks -- scripts/common.sh@334 -- # local ver2 ver2_l 00:07:35.346 15:13:25 event.cpu_locks -- scripts/common.sh@336 -- # IFS=.-: 00:07:35.346 15:13:25 event.cpu_locks -- scripts/common.sh@336 -- # read -ra ver1 00:07:35.346 15:13:25 event.cpu_locks -- scripts/common.sh@337 -- # IFS=.-: 00:07:35.346 15:13:25 event.cpu_locks -- scripts/common.sh@337 -- # read -ra ver2 00:07:35.346 15:13:25 event.cpu_locks -- scripts/common.sh@338 -- # local 'op=<' 00:07:35.346 15:13:25 event.cpu_locks -- scripts/common.sh@340 -- # ver1_l=2 00:07:35.346 15:13:25 event.cpu_locks -- scripts/common.sh@341 -- # ver2_l=1 00:07:35.346 15:13:25 event.cpu_locks -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:07:35.346 15:13:25 event.cpu_locks -- scripts/common.sh@344 -- # case "$op" in 00:07:35.346 15:13:25 event.cpu_locks -- scripts/common.sh@345 -- # : 1 00:07:35.346 15:13:25 event.cpu_locks -- scripts/common.sh@364 -- # (( v = 0 )) 00:07:35.346 15:13:25 event.cpu_locks -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:07:35.346 15:13:25 event.cpu_locks -- scripts/common.sh@365 -- # decimal 1 00:07:35.346 15:13:25 event.cpu_locks -- scripts/common.sh@353 -- # local d=1 00:07:35.346 15:13:25 event.cpu_locks -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:07:35.346 15:13:25 event.cpu_locks -- scripts/common.sh@355 -- # echo 1 00:07:35.346 15:13:25 event.cpu_locks -- scripts/common.sh@365 -- # ver1[v]=1 00:07:35.346 15:13:25 event.cpu_locks -- scripts/common.sh@366 -- # decimal 2 00:07:35.346 15:13:25 event.cpu_locks -- scripts/common.sh@353 -- # local d=2 00:07:35.346 15:13:25 event.cpu_locks -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:07:35.346 15:13:25 event.cpu_locks -- scripts/common.sh@355 -- # echo 2 00:07:35.346 15:13:25 event.cpu_locks -- scripts/common.sh@366 -- # ver2[v]=2 00:07:35.346 15:13:25 event.cpu_locks -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:07:35.346 15:13:25 event.cpu_locks -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:07:35.346 15:13:25 event.cpu_locks -- scripts/common.sh@368 -- # return 0 00:07:35.346 15:13:25 event.cpu_locks -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:07:35.346 15:13:25 event.cpu_locks -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:07:35.346 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:35.346 --rc genhtml_branch_coverage=1 00:07:35.346 --rc genhtml_function_coverage=1 00:07:35.346 --rc genhtml_legend=1 00:07:35.346 --rc geninfo_all_blocks=1 00:07:35.346 --rc geninfo_unexecuted_blocks=1 00:07:35.346 00:07:35.346 ' 00:07:35.346 15:13:25 event.cpu_locks -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:07:35.346 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:35.346 --rc genhtml_branch_coverage=1 00:07:35.346 --rc genhtml_function_coverage=1 00:07:35.346 --rc genhtml_legend=1 00:07:35.346 --rc geninfo_all_blocks=1 00:07:35.346 --rc geninfo_unexecuted_blocks=1 00:07:35.346 00:07:35.346 ' 00:07:35.346 15:13:25 event.cpu_locks -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:07:35.346 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:35.346 --rc genhtml_branch_coverage=1 00:07:35.346 --rc genhtml_function_coverage=1 00:07:35.346 --rc genhtml_legend=1 00:07:35.346 --rc geninfo_all_blocks=1 00:07:35.346 --rc geninfo_unexecuted_blocks=1 00:07:35.346 00:07:35.346 ' 00:07:35.346 15:13:25 event.cpu_locks -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:07:35.346 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:35.346 --rc genhtml_branch_coverage=1 00:07:35.346 --rc genhtml_function_coverage=1 00:07:35.346 --rc genhtml_legend=1 00:07:35.346 --rc geninfo_all_blocks=1 00:07:35.346 --rc geninfo_unexecuted_blocks=1 00:07:35.346 00:07:35.346 ' 00:07:35.346 15:13:25 event.cpu_locks -- event/cpu_locks.sh@11 -- # rpc_sock1=/var/tmp/spdk.sock 00:07:35.346 15:13:25 event.cpu_locks -- event/cpu_locks.sh@12 -- # rpc_sock2=/var/tmp/spdk2.sock 00:07:35.346 15:13:25 event.cpu_locks -- event/cpu_locks.sh@164 -- # trap cleanup EXIT SIGTERM SIGINT 00:07:35.346 15:13:25 event.cpu_locks -- event/cpu_locks.sh@166 -- # run_test default_locks default_locks 00:07:35.346 15:13:25 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:07:35.346 15:13:25 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:35.346 15:13:25 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:07:35.346 ************************************ 00:07:35.346 START TEST default_locks 00:07:35.346 ************************************ 00:07:35.346 15:13:25 event.cpu_locks.default_locks -- common/autotest_common.sh@1129 -- # default_locks 00:07:35.346 15:13:25 event.cpu_locks.default_locks -- event/cpu_locks.sh@46 -- # spdk_tgt_pid=70696 00:07:35.346 15:13:25 event.cpu_locks.default_locks -- event/cpu_locks.sh@45 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:07:35.346 15:13:25 event.cpu_locks.default_locks -- event/cpu_locks.sh@47 -- # waitforlisten 70696 00:07:35.346 15:13:25 event.cpu_locks.default_locks -- common/autotest_common.sh@835 -- # '[' -z 70696 ']' 00:07:35.346 15:13:25 event.cpu_locks.default_locks -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:35.346 15:13:25 event.cpu_locks.default_locks -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:35.346 15:13:25 event.cpu_locks.default_locks -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:35.346 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:35.346 15:13:25 event.cpu_locks.default_locks -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:35.346 15:13:25 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:07:35.606 [2024-11-19 15:13:25.707624] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:07:35.606 [2024-11-19 15:13:25.707790] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70696 ] 00:07:35.606 [2024-11-19 15:13:25.861146] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:35.606 [2024-11-19 15:13:25.900378] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:36.542 15:13:26 event.cpu_locks.default_locks -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:36.542 15:13:26 event.cpu_locks.default_locks -- common/autotest_common.sh@868 -- # return 0 00:07:36.542 15:13:26 event.cpu_locks.default_locks -- event/cpu_locks.sh@49 -- # locks_exist 70696 00:07:36.542 15:13:26 event.cpu_locks.default_locks -- event/cpu_locks.sh@22 -- # lslocks -p 70696 00:07:36.542 15:13:26 event.cpu_locks.default_locks -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:07:36.542 15:13:26 event.cpu_locks.default_locks -- event/cpu_locks.sh@50 -- # killprocess 70696 00:07:36.542 15:13:26 event.cpu_locks.default_locks -- common/autotest_common.sh@954 -- # '[' -z 70696 ']' 00:07:36.542 15:13:26 event.cpu_locks.default_locks -- common/autotest_common.sh@958 -- # kill -0 70696 00:07:36.542 15:13:26 event.cpu_locks.default_locks -- common/autotest_common.sh@959 -- # uname 00:07:36.542 15:13:26 event.cpu_locks.default_locks -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:36.542 15:13:26 event.cpu_locks.default_locks -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70696 00:07:36.542 15:13:26 event.cpu_locks.default_locks -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:36.542 15:13:26 event.cpu_locks.default_locks -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:36.542 killing process with pid 70696 00:07:36.542 15:13:26 event.cpu_locks.default_locks -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70696' 00:07:36.542 15:13:26 event.cpu_locks.default_locks -- common/autotest_common.sh@973 -- # kill 70696 00:07:36.542 15:13:26 event.cpu_locks.default_locks -- common/autotest_common.sh@978 -- # wait 70696 00:07:37.112 15:13:27 event.cpu_locks.default_locks -- event/cpu_locks.sh@52 -- # NOT waitforlisten 70696 00:07:37.112 15:13:27 event.cpu_locks.default_locks -- common/autotest_common.sh@652 -- # local es=0 00:07:37.112 15:13:27 event.cpu_locks.default_locks -- common/autotest_common.sh@654 -- # valid_exec_arg waitforlisten 70696 00:07:37.112 15:13:27 event.cpu_locks.default_locks -- common/autotest_common.sh@640 -- # local arg=waitforlisten 00:07:37.112 15:13:27 event.cpu_locks.default_locks -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:07:37.112 15:13:27 event.cpu_locks.default_locks -- common/autotest_common.sh@644 -- # type -t waitforlisten 00:07:37.112 15:13:27 event.cpu_locks.default_locks -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:07:37.112 15:13:27 event.cpu_locks.default_locks -- common/autotest_common.sh@655 -- # waitforlisten 70696 00:07:37.112 15:13:27 event.cpu_locks.default_locks -- common/autotest_common.sh@835 -- # '[' -z 70696 ']' 00:07:37.112 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:37.112 ERROR: process (pid: 70696) is no longer running 00:07:37.112 15:13:27 event.cpu_locks.default_locks -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:37.112 15:13:27 event.cpu_locks.default_locks -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:37.112 15:13:27 event.cpu_locks.default_locks -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:37.112 15:13:27 event.cpu_locks.default_locks -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:37.112 15:13:27 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:07:37.112 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 850: kill: (70696) - No such process 00:07:37.112 15:13:27 event.cpu_locks.default_locks -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:37.112 15:13:27 event.cpu_locks.default_locks -- common/autotest_common.sh@868 -- # return 1 00:07:37.112 15:13:27 event.cpu_locks.default_locks -- common/autotest_common.sh@655 -- # es=1 00:07:37.112 15:13:27 event.cpu_locks.default_locks -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:07:37.112 15:13:27 event.cpu_locks.default_locks -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:07:37.112 15:13:27 event.cpu_locks.default_locks -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:07:37.112 15:13:27 event.cpu_locks.default_locks -- event/cpu_locks.sh@54 -- # no_locks 00:07:37.112 15:13:27 event.cpu_locks.default_locks -- event/cpu_locks.sh@26 -- # lock_files=() 00:07:37.112 15:13:27 event.cpu_locks.default_locks -- event/cpu_locks.sh@26 -- # local lock_files 00:07:37.112 15:13:27 event.cpu_locks.default_locks -- event/cpu_locks.sh@27 -- # (( 0 != 0 )) 00:07:37.112 00:07:37.112 real 0m1.681s 00:07:37.112 user 0m1.507s 00:07:37.112 sys 0m0.609s 00:07:37.112 15:13:27 event.cpu_locks.default_locks -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:37.112 15:13:27 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:07:37.112 ************************************ 00:07:37.112 END TEST default_locks 00:07:37.112 ************************************ 00:07:37.112 15:13:27 event.cpu_locks -- event/cpu_locks.sh@167 -- # run_test default_locks_via_rpc default_locks_via_rpc 00:07:37.112 15:13:27 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:07:37.112 15:13:27 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:37.112 15:13:27 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:07:37.112 ************************************ 00:07:37.112 START TEST default_locks_via_rpc 00:07:37.112 ************************************ 00:07:37.112 15:13:27 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@1129 -- # default_locks_via_rpc 00:07:37.112 15:13:27 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@62 -- # spdk_tgt_pid=70745 00:07:37.112 15:13:27 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@61 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:07:37.112 15:13:27 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@63 -- # waitforlisten 70745 00:07:37.112 15:13:27 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 70745 ']' 00:07:37.112 15:13:27 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:37.112 15:13:27 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:37.112 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:37.112 15:13:27 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:37.113 15:13:27 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:37.113 15:13:27 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:07:37.372 [2024-11-19 15:13:27.478136] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:07:37.372 [2024-11-19 15:13:27.478273] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70745 ] 00:07:37.372 [2024-11-19 15:13:27.630217] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:37.372 [2024-11-19 15:13:27.669891] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:37.941 15:13:28 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:37.941 15:13:28 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:07:37.941 15:13:28 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@65 -- # rpc_cmd framework_disable_cpumask_locks 00:07:37.941 15:13:28 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:37.941 15:13:28 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:07:37.941 15:13:28 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:37.941 15:13:28 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@67 -- # no_locks 00:07:37.941 15:13:28 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@26 -- # lock_files=() 00:07:37.941 15:13:28 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@26 -- # local lock_files 00:07:37.942 15:13:28 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@27 -- # (( 0 != 0 )) 00:07:37.942 15:13:28 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@69 -- # rpc_cmd framework_enable_cpumask_locks 00:07:37.942 15:13:28 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:37.942 15:13:28 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:07:38.201 15:13:28 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:38.201 15:13:28 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@71 -- # locks_exist 70745 00:07:38.201 15:13:28 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:07:38.201 15:13:28 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@22 -- # lslocks -p 70745 00:07:38.461 15:13:28 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@73 -- # killprocess 70745 00:07:38.461 15:13:28 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@954 -- # '[' -z 70745 ']' 00:07:38.461 15:13:28 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@958 -- # kill -0 70745 00:07:38.461 15:13:28 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@959 -- # uname 00:07:38.461 15:13:28 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:38.461 15:13:28 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70745 00:07:38.461 15:13:28 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:38.461 15:13:28 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:38.461 15:13:28 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70745' 00:07:38.461 killing process with pid 70745 00:07:38.461 15:13:28 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@973 -- # kill 70745 00:07:38.461 15:13:28 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@978 -- # wait 70745 00:07:39.035 00:07:39.035 real 0m1.943s 00:07:39.035 user 0m1.762s 00:07:39.035 sys 0m0.735s 00:07:39.035 15:13:29 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:39.035 15:13:29 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:07:39.035 ************************************ 00:07:39.035 END TEST default_locks_via_rpc 00:07:39.035 ************************************ 00:07:39.036 15:13:29 event.cpu_locks -- event/cpu_locks.sh@168 -- # run_test non_locking_app_on_locked_coremask non_locking_app_on_locked_coremask 00:07:39.036 15:13:29 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:07:39.036 15:13:29 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:39.036 15:13:29 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:07:39.298 ************************************ 00:07:39.298 START TEST non_locking_app_on_locked_coremask 00:07:39.298 ************************************ 00:07:39.298 15:13:29 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@1129 -- # non_locking_app_on_locked_coremask 00:07:39.298 15:13:29 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@80 -- # spdk_tgt_pid=70797 00:07:39.298 15:13:29 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@79 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:07:39.298 15:13:29 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@81 -- # waitforlisten 70797 /var/tmp/spdk.sock 00:07:39.298 15:13:29 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # '[' -z 70797 ']' 00:07:39.298 15:13:29 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:39.298 15:13:29 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:39.298 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:39.298 15:13:29 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:39.298 15:13:29 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:39.298 15:13:29 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:07:39.298 [2024-11-19 15:13:29.489925] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:07:39.298 [2024-11-19 15:13:29.490078] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70797 ] 00:07:39.558 [2024-11-19 15:13:29.638786] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:39.558 [2024-11-19 15:13:29.679375] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:40.170 15:13:30 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:40.170 15:13:30 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@868 -- # return 0 00:07:40.170 15:13:30 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@84 -- # spdk_tgt_pid2=70813 00:07:40.170 15:13:30 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@85 -- # waitforlisten 70813 /var/tmp/spdk2.sock 00:07:40.170 15:13:30 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@83 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 --disable-cpumask-locks -r /var/tmp/spdk2.sock 00:07:40.170 15:13:30 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # '[' -z 70813 ']' 00:07:40.170 15:13:30 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:07:40.170 15:13:30 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:40.170 15:13:30 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:07:40.170 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:07:40.170 15:13:30 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:40.170 15:13:30 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:07:40.170 [2024-11-19 15:13:30.384228] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:07:40.170 [2024-11-19 15:13:30.384397] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70813 ] 00:07:40.430 [2024-11-19 15:13:30.538683] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:07:40.430 [2024-11-19 15:13:30.538743] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:40.430 [2024-11-19 15:13:30.624517] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:41.000 15:13:31 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:41.000 15:13:31 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@868 -- # return 0 00:07:41.000 15:13:31 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@87 -- # locks_exist 70797 00:07:41.000 15:13:31 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 70797 00:07:41.000 15:13:31 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:07:41.571 15:13:31 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@89 -- # killprocess 70797 00:07:41.571 15:13:31 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # '[' -z 70797 ']' 00:07:41.571 15:13:31 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@958 -- # kill -0 70797 00:07:41.571 15:13:31 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # uname 00:07:41.571 15:13:31 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:41.571 15:13:31 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70797 00:07:41.571 15:13:31 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:41.571 15:13:31 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:41.571 killing process with pid 70797 00:07:41.571 15:13:31 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70797' 00:07:41.571 15:13:31 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@973 -- # kill 70797 00:07:41.571 15:13:31 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@978 -- # wait 70797 00:07:42.953 15:13:33 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@90 -- # killprocess 70813 00:07:42.953 15:13:33 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # '[' -z 70813 ']' 00:07:42.953 15:13:33 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@958 -- # kill -0 70813 00:07:42.953 15:13:33 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # uname 00:07:42.953 15:13:33 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:42.953 15:13:33 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70813 00:07:42.953 15:13:33 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:42.953 15:13:33 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:42.953 killing process with pid 70813 00:07:42.953 15:13:33 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70813' 00:07:42.953 15:13:33 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@973 -- # kill 70813 00:07:42.953 15:13:33 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@978 -- # wait 70813 00:07:43.523 00:07:43.523 real 0m4.281s 00:07:43.523 user 0m4.136s 00:07:43.523 sys 0m1.373s 00:07:43.523 15:13:33 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:43.523 ************************************ 00:07:43.523 END TEST non_locking_app_on_locked_coremask 00:07:43.523 ************************************ 00:07:43.523 15:13:33 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:07:43.523 15:13:33 event.cpu_locks -- event/cpu_locks.sh@169 -- # run_test locking_app_on_unlocked_coremask locking_app_on_unlocked_coremask 00:07:43.523 15:13:33 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:07:43.523 15:13:33 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:43.523 15:13:33 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:07:43.523 ************************************ 00:07:43.523 START TEST locking_app_on_unlocked_coremask 00:07:43.523 ************************************ 00:07:43.523 15:13:33 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@1129 -- # locking_app_on_unlocked_coremask 00:07:43.523 15:13:33 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@98 -- # spdk_tgt_pid=70882 00:07:43.523 15:13:33 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@97 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 --disable-cpumask-locks 00:07:43.523 15:13:33 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@99 -- # waitforlisten 70882 /var/tmp/spdk.sock 00:07:43.523 15:13:33 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@835 -- # '[' -z 70882 ']' 00:07:43.523 15:13:33 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:43.523 15:13:33 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:43.523 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:43.523 15:13:33 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:43.523 15:13:33 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:43.523 15:13:33 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:07:43.523 [2024-11-19 15:13:33.831818] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:07:43.523 [2024-11-19 15:13:33.831959] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70882 ] 00:07:43.783 [2024-11-19 15:13:33.965541] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:07:43.783 [2024-11-19 15:13:33.965591] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:43.783 [2024-11-19 15:13:34.005799] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:44.352 15:13:34 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:44.352 15:13:34 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@868 -- # return 0 00:07:44.352 15:13:34 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@102 -- # spdk_tgt_pid2=70898 00:07:44.352 15:13:34 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@101 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 -r /var/tmp/spdk2.sock 00:07:44.352 15:13:34 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@103 -- # waitforlisten 70898 /var/tmp/spdk2.sock 00:07:44.352 15:13:34 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@835 -- # '[' -z 70898 ']' 00:07:44.352 15:13:34 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:07:44.352 15:13:34 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:44.352 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:07:44.352 15:13:34 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:07:44.352 15:13:34 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:44.353 15:13:34 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:07:44.612 [2024-11-19 15:13:34.741243] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:07:44.612 [2024-11-19 15:13:34.741372] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70898 ] 00:07:44.612 [2024-11-19 15:13:34.899295] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:44.871 [2024-11-19 15:13:34.985373] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:45.439 15:13:35 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:45.439 15:13:35 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@868 -- # return 0 00:07:45.439 15:13:35 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@105 -- # locks_exist 70898 00:07:45.439 15:13:35 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 70898 00:07:45.439 15:13:35 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:07:46.007 15:13:36 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@107 -- # killprocess 70882 00:07:46.007 15:13:36 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@954 -- # '[' -z 70882 ']' 00:07:46.007 15:13:36 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@958 -- # kill -0 70882 00:07:46.007 15:13:36 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@959 -- # uname 00:07:46.007 15:13:36 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:46.007 15:13:36 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70882 00:07:46.007 15:13:36 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:46.007 15:13:36 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:46.007 killing process with pid 70882 00:07:46.007 15:13:36 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70882' 00:07:46.007 15:13:36 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@973 -- # kill 70882 00:07:46.007 15:13:36 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@978 -- # wait 70882 00:07:47.386 15:13:37 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@108 -- # killprocess 70898 00:07:47.386 15:13:37 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@954 -- # '[' -z 70898 ']' 00:07:47.386 15:13:37 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@958 -- # kill -0 70898 00:07:47.386 15:13:37 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@959 -- # uname 00:07:47.386 15:13:37 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:47.386 15:13:37 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70898 00:07:47.386 15:13:37 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:47.386 15:13:37 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:47.386 15:13:37 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70898' 00:07:47.386 killing process with pid 70898 00:07:47.386 15:13:37 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@973 -- # kill 70898 00:07:47.386 15:13:37 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@978 -- # wait 70898 00:07:47.646 00:07:47.646 real 0m4.186s 00:07:47.646 user 0m4.088s 00:07:47.646 sys 0m1.270s 00:07:47.646 15:13:37 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:47.646 15:13:37 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:07:47.646 ************************************ 00:07:47.646 END TEST locking_app_on_unlocked_coremask 00:07:47.646 ************************************ 00:07:47.646 15:13:37 event.cpu_locks -- event/cpu_locks.sh@170 -- # run_test locking_app_on_locked_coremask locking_app_on_locked_coremask 00:07:47.646 15:13:37 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:07:47.646 15:13:37 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:47.646 15:13:37 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:07:47.926 ************************************ 00:07:47.926 START TEST locking_app_on_locked_coremask 00:07:47.926 ************************************ 00:07:47.926 15:13:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@1129 -- # locking_app_on_locked_coremask 00:07:47.926 15:13:37 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@115 -- # spdk_tgt_pid=70969 00:07:47.926 15:13:37 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@114 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:07:47.926 15:13:37 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@116 -- # waitforlisten 70969 /var/tmp/spdk.sock 00:07:47.926 15:13:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # '[' -z 70969 ']' 00:07:47.926 15:13:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:47.926 15:13:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:47.926 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:47.926 15:13:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:47.926 15:13:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:47.926 15:13:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:07:47.926 [2024-11-19 15:13:38.098472] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:07:47.927 [2024-11-19 15:13:38.098619] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70969 ] 00:07:48.200 [2024-11-19 15:13:38.256132] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:48.200 [2024-11-19 15:13:38.297117] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:48.768 15:13:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:48.768 15:13:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@868 -- # return 0 00:07:48.768 15:13:38 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@119 -- # spdk_tgt_pid2=70985 00:07:48.768 15:13:38 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@118 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 -r /var/tmp/spdk2.sock 00:07:48.768 15:13:38 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@120 -- # NOT waitforlisten 70985 /var/tmp/spdk2.sock 00:07:48.768 15:13:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@652 -- # local es=0 00:07:48.768 15:13:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@654 -- # valid_exec_arg waitforlisten 70985 /var/tmp/spdk2.sock 00:07:48.768 15:13:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@640 -- # local arg=waitforlisten 00:07:48.768 15:13:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:07:48.768 15:13:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@644 -- # type -t waitforlisten 00:07:48.768 15:13:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:07:48.768 15:13:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@655 -- # waitforlisten 70985 /var/tmp/spdk2.sock 00:07:48.768 15:13:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # '[' -z 70985 ']' 00:07:48.768 15:13:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:07:48.768 15:13:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:48.768 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:07:48.768 15:13:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:07:48.768 15:13:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:48.768 15:13:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:07:48.768 [2024-11-19 15:13:39.006507] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:07:48.768 [2024-11-19 15:13:39.006638] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70985 ] 00:07:49.038 [2024-11-19 15:13:39.157360] app.c: 781:claim_cpu_cores: *ERROR*: Cannot create lock on core 0, probably process 70969 has claimed it. 00:07:49.038 [2024-11-19 15:13:39.157437] app.c: 912:spdk_app_start: *ERROR*: Unable to acquire lock on assigned core mask - exiting. 00:07:49.304 ERROR: process (pid: 70985) is no longer running 00:07:49.304 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 850: kill: (70985) - No such process 00:07:49.304 15:13:39 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:49.304 15:13:39 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@868 -- # return 1 00:07:49.304 15:13:39 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@655 -- # es=1 00:07:49.304 15:13:39 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:07:49.304 15:13:39 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:07:49.304 15:13:39 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:07:49.304 15:13:39 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@122 -- # locks_exist 70969 00:07:49.304 15:13:39 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 70969 00:07:49.304 15:13:39 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:07:49.874 15:13:40 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@124 -- # killprocess 70969 00:07:49.874 15:13:40 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # '[' -z 70969 ']' 00:07:49.874 15:13:40 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@958 -- # kill -0 70969 00:07:49.874 15:13:40 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # uname 00:07:49.874 15:13:40 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:49.874 15:13:40 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70969 00:07:49.874 15:13:40 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:49.874 15:13:40 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:49.874 killing process with pid 70969 00:07:49.874 15:13:40 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70969' 00:07:49.874 15:13:40 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@973 -- # kill 70969 00:07:49.874 15:13:40 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@978 -- # wait 70969 00:07:50.443 00:07:50.443 real 0m2.669s 00:07:50.443 user 0m2.731s 00:07:50.443 sys 0m0.857s 00:07:50.443 15:13:40 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:50.443 15:13:40 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:07:50.443 ************************************ 00:07:50.443 END TEST locking_app_on_locked_coremask 00:07:50.443 ************************************ 00:07:50.443 15:13:40 event.cpu_locks -- event/cpu_locks.sh@171 -- # run_test locking_overlapped_coremask locking_overlapped_coremask 00:07:50.443 15:13:40 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:07:50.443 15:13:40 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:50.443 15:13:40 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:07:50.443 ************************************ 00:07:50.443 START TEST locking_overlapped_coremask 00:07:50.443 ************************************ 00:07:50.444 15:13:40 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@1129 -- # locking_overlapped_coremask 00:07:50.444 15:13:40 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@132 -- # spdk_tgt_pid=71038 00:07:50.444 15:13:40 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@131 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x7 00:07:50.444 15:13:40 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@133 -- # waitforlisten 71038 /var/tmp/spdk.sock 00:07:50.444 15:13:40 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@835 -- # '[' -z 71038 ']' 00:07:50.444 15:13:40 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:50.444 15:13:40 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:50.444 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:50.444 15:13:40 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:50.444 15:13:40 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:50.444 15:13:40 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:07:50.703 [2024-11-19 15:13:40.829914] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:07:50.703 [2024-11-19 15:13:40.830047] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71038 ] 00:07:50.703 [2024-11-19 15:13:40.981993] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:07:50.703 [2024-11-19 15:13:41.025731] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:07:50.703 [2024-11-19 15:13:41.025835] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:50.703 [2024-11-19 15:13:41.026018] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:07:51.644 15:13:41 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:51.644 15:13:41 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@868 -- # return 0 00:07:51.644 15:13:41 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@136 -- # spdk_tgt_pid2=71056 00:07:51.644 15:13:41 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@135 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1c -r /var/tmp/spdk2.sock 00:07:51.644 15:13:41 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@137 -- # NOT waitforlisten 71056 /var/tmp/spdk2.sock 00:07:51.644 15:13:41 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@652 -- # local es=0 00:07:51.644 15:13:41 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@654 -- # valid_exec_arg waitforlisten 71056 /var/tmp/spdk2.sock 00:07:51.644 15:13:41 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@640 -- # local arg=waitforlisten 00:07:51.644 15:13:41 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:07:51.644 15:13:41 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@644 -- # type -t waitforlisten 00:07:51.644 15:13:41 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:07:51.644 15:13:41 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@655 -- # waitforlisten 71056 /var/tmp/spdk2.sock 00:07:51.644 15:13:41 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@835 -- # '[' -z 71056 ']' 00:07:51.644 15:13:41 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:07:51.644 15:13:41 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:51.644 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:07:51.644 15:13:41 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:07:51.644 15:13:41 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:51.644 15:13:41 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:07:51.644 [2024-11-19 15:13:41.731434] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:07:51.644 [2024-11-19 15:13:41.731585] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1c --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71056 ] 00:07:51.644 [2024-11-19 15:13:41.884277] app.c: 781:claim_cpu_cores: *ERROR*: Cannot create lock on core 2, probably process 71038 has claimed it. 00:07:51.644 [2024-11-19 15:13:41.884349] app.c: 912:spdk_app_start: *ERROR*: Unable to acquire lock on assigned core mask - exiting. 00:07:52.214 ERROR: process (pid: 71056) is no longer running 00:07:52.214 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 850: kill: (71056) - No such process 00:07:52.214 15:13:42 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:52.214 15:13:42 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@868 -- # return 1 00:07:52.214 15:13:42 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@655 -- # es=1 00:07:52.214 15:13:42 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:07:52.214 15:13:42 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:07:52.214 15:13:42 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:07:52.214 15:13:42 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@139 -- # check_remaining_locks 00:07:52.214 15:13:42 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@36 -- # locks=(/var/tmp/spdk_cpu_lock_*) 00:07:52.214 15:13:42 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@37 -- # locks_expected=(/var/tmp/spdk_cpu_lock_{000..002}) 00:07:52.214 15:13:42 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@38 -- # [[ /var/tmp/spdk_cpu_lock_000 /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 == \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\0\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\1\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\2 ]] 00:07:52.214 15:13:42 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@141 -- # killprocess 71038 00:07:52.214 15:13:42 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@954 -- # '[' -z 71038 ']' 00:07:52.214 15:13:42 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@958 -- # kill -0 71038 00:07:52.214 15:13:42 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@959 -- # uname 00:07:52.214 15:13:42 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:52.214 15:13:42 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71038 00:07:52.214 15:13:42 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:52.214 15:13:42 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:52.214 killing process with pid 71038 00:07:52.214 15:13:42 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71038' 00:07:52.214 15:13:42 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@973 -- # kill 71038 00:07:52.214 15:13:42 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@978 -- # wait 71038 00:07:52.785 00:07:52.785 real 0m2.284s 00:07:52.785 user 0m5.976s 00:07:52.785 sys 0m0.660s 00:07:52.785 15:13:43 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:52.785 15:13:43 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:07:52.785 ************************************ 00:07:52.785 END TEST locking_overlapped_coremask 00:07:52.785 ************************************ 00:07:52.785 15:13:43 event.cpu_locks -- event/cpu_locks.sh@172 -- # run_test locking_overlapped_coremask_via_rpc locking_overlapped_coremask_via_rpc 00:07:52.785 15:13:43 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:07:52.785 15:13:43 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:52.785 15:13:43 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:07:52.785 ************************************ 00:07:52.785 START TEST locking_overlapped_coremask_via_rpc 00:07:52.785 ************************************ 00:07:52.785 15:13:43 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@1129 -- # locking_overlapped_coremask_via_rpc 00:07:52.785 15:13:43 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@148 -- # spdk_tgt_pid=71098 00:07:52.785 15:13:43 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@147 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x7 --disable-cpumask-locks 00:07:52.785 15:13:43 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@149 -- # waitforlisten 71098 /var/tmp/spdk.sock 00:07:52.785 15:13:43 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 71098 ']' 00:07:52.785 15:13:43 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:52.785 15:13:43 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:52.785 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:52.785 15:13:43 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:52.785 15:13:43 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:52.785 15:13:43 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:07:53.046 [2024-11-19 15:13:43.189941] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:07:53.046 [2024-11-19 15:13:43.190096] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71098 ] 00:07:53.046 [2024-11-19 15:13:43.348348] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:07:53.046 [2024-11-19 15:13:43.348399] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:07:53.306 [2024-11-19 15:13:43.391377] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:07:53.306 [2024-11-19 15:13:43.391485] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:53.306 [2024-11-19 15:13:43.391640] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:07:53.876 15:13:43 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:53.876 15:13:43 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:07:53.876 15:13:43 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@151 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1c -r /var/tmp/spdk2.sock --disable-cpumask-locks 00:07:53.876 15:13:43 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@152 -- # spdk_tgt_pid2=71116 00:07:53.876 15:13:43 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@153 -- # waitforlisten 71116 /var/tmp/spdk2.sock 00:07:53.876 15:13:43 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 71116 ']' 00:07:53.876 15:13:43 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:07:53.876 15:13:43 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:53.876 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:07:53.876 15:13:43 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:07:53.876 15:13:44 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:53.876 15:13:44 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:07:53.876 [2024-11-19 15:13:44.081632] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:07:53.876 [2024-11-19 15:13:44.081799] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1c --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71116 ] 00:07:54.136 [2024-11-19 15:13:44.238158] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:07:54.136 [2024-11-19 15:13:44.238218] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:07:54.136 [2024-11-19 15:13:44.335592] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:07:54.136 [2024-11-19 15:13:44.335767] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:07:54.136 [2024-11-19 15:13:44.335508] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:07:54.742 15:13:44 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:54.742 15:13:44 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:07:54.742 15:13:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@155 -- # rpc_cmd framework_enable_cpumask_locks 00:07:54.742 15:13:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:54.742 15:13:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:07:54.742 15:13:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:54.742 15:13:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@156 -- # NOT rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:07:54.742 15:13:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@652 -- # local es=0 00:07:54.742 15:13:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:07:54.742 15:13:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:07:54.742 15:13:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:07:54.742 15:13:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:07:54.742 15:13:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:07:54.742 15:13:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@655 -- # rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:07:54.742 15:13:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:54.742 15:13:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:07:54.742 [2024-11-19 15:13:45.018157] app.c: 781:claim_cpu_cores: *ERROR*: Cannot create lock on core 2, probably process 71098 has claimed it. 00:07:54.742 request: 00:07:54.742 { 00:07:54.742 "method": "framework_enable_cpumask_locks", 00:07:54.742 "req_id": 1 00:07:54.742 } 00:07:54.742 Got JSON-RPC error response 00:07:54.742 response: 00:07:54.742 { 00:07:54.742 "code": -32603, 00:07:54.742 "message": "Failed to claim CPU core: 2" 00:07:54.742 } 00:07:54.742 15:13:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:07:54.742 15:13:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@655 -- # es=1 00:07:54.742 15:13:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:07:54.742 15:13:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:07:54.742 15:13:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:07:54.742 15:13:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@158 -- # waitforlisten 71098 /var/tmp/spdk.sock 00:07:54.742 15:13:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 71098 ']' 00:07:54.742 15:13:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:54.742 15:13:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:54.742 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:54.742 15:13:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:54.742 15:13:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:54.742 15:13:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:07:55.017 15:13:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:55.017 15:13:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:07:55.017 15:13:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@159 -- # waitforlisten 71116 /var/tmp/spdk2.sock 00:07:55.017 15:13:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 71116 ']' 00:07:55.017 15:13:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:07:55.017 15:13:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:55.017 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:07:55.017 15:13:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:07:55.017 15:13:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:55.017 15:13:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:07:55.277 15:13:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:55.277 15:13:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:07:55.277 15:13:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@161 -- # check_remaining_locks 00:07:55.277 15:13:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@36 -- # locks=(/var/tmp/spdk_cpu_lock_*) 00:07:55.277 15:13:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@37 -- # locks_expected=(/var/tmp/spdk_cpu_lock_{000..002}) 00:07:55.277 15:13:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@38 -- # [[ /var/tmp/spdk_cpu_lock_000 /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 == \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\0\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\1\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\2 ]] 00:07:55.277 00:07:55.277 real 0m2.372s 00:07:55.277 user 0m1.061s 00:07:55.277 sys 0m0.183s 00:07:55.277 15:13:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:55.277 15:13:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:07:55.277 ************************************ 00:07:55.277 END TEST locking_overlapped_coremask_via_rpc 00:07:55.277 ************************************ 00:07:55.277 15:13:45 event.cpu_locks -- event/cpu_locks.sh@174 -- # cleanup 00:07:55.277 15:13:45 event.cpu_locks -- event/cpu_locks.sh@15 -- # [[ -z 71098 ]] 00:07:55.277 15:13:45 event.cpu_locks -- event/cpu_locks.sh@15 -- # killprocess 71098 00:07:55.278 15:13:45 event.cpu_locks -- common/autotest_common.sh@954 -- # '[' -z 71098 ']' 00:07:55.278 15:13:45 event.cpu_locks -- common/autotest_common.sh@958 -- # kill -0 71098 00:07:55.278 15:13:45 event.cpu_locks -- common/autotest_common.sh@959 -- # uname 00:07:55.278 15:13:45 event.cpu_locks -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:55.278 15:13:45 event.cpu_locks -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71098 00:07:55.278 15:13:45 event.cpu_locks -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:55.278 15:13:45 event.cpu_locks -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:55.278 killing process with pid 71098 00:07:55.278 15:13:45 event.cpu_locks -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71098' 00:07:55.278 15:13:45 event.cpu_locks -- common/autotest_common.sh@973 -- # kill 71098 00:07:55.278 15:13:45 event.cpu_locks -- common/autotest_common.sh@978 -- # wait 71098 00:07:55.847 15:13:46 event.cpu_locks -- event/cpu_locks.sh@16 -- # [[ -z 71116 ]] 00:07:55.847 15:13:46 event.cpu_locks -- event/cpu_locks.sh@16 -- # killprocess 71116 00:07:55.847 15:13:46 event.cpu_locks -- common/autotest_common.sh@954 -- # '[' -z 71116 ']' 00:07:55.847 15:13:46 event.cpu_locks -- common/autotest_common.sh@958 -- # kill -0 71116 00:07:55.847 15:13:46 event.cpu_locks -- common/autotest_common.sh@959 -- # uname 00:07:55.847 15:13:46 event.cpu_locks -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:55.848 15:13:46 event.cpu_locks -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71116 00:07:56.108 killing process with pid 71116 00:07:56.108 15:13:46 event.cpu_locks -- common/autotest_common.sh@960 -- # process_name=reactor_2 00:07:56.108 15:13:46 event.cpu_locks -- common/autotest_common.sh@964 -- # '[' reactor_2 = sudo ']' 00:07:56.108 15:13:46 event.cpu_locks -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71116' 00:07:56.108 15:13:46 event.cpu_locks -- common/autotest_common.sh@973 -- # kill 71116 00:07:56.108 15:13:46 event.cpu_locks -- common/autotest_common.sh@978 -- # wait 71116 00:07:56.678 15:13:46 event.cpu_locks -- event/cpu_locks.sh@18 -- # rm -f 00:07:56.678 Process with pid 71098 is not found 00:07:56.678 Process with pid 71116 is not found 00:07:56.678 15:13:46 event.cpu_locks -- event/cpu_locks.sh@1 -- # cleanup 00:07:56.678 15:13:46 event.cpu_locks -- event/cpu_locks.sh@15 -- # [[ -z 71098 ]] 00:07:56.678 15:13:46 event.cpu_locks -- event/cpu_locks.sh@15 -- # killprocess 71098 00:07:56.678 15:13:46 event.cpu_locks -- common/autotest_common.sh@954 -- # '[' -z 71098 ']' 00:07:56.678 15:13:46 event.cpu_locks -- common/autotest_common.sh@958 -- # kill -0 71098 00:07:56.678 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 958: kill: (71098) - No such process 00:07:56.678 15:13:46 event.cpu_locks -- common/autotest_common.sh@981 -- # echo 'Process with pid 71098 is not found' 00:07:56.678 15:13:46 event.cpu_locks -- event/cpu_locks.sh@16 -- # [[ -z 71116 ]] 00:07:56.678 15:13:46 event.cpu_locks -- event/cpu_locks.sh@16 -- # killprocess 71116 00:07:56.678 15:13:46 event.cpu_locks -- common/autotest_common.sh@954 -- # '[' -z 71116 ']' 00:07:56.678 15:13:46 event.cpu_locks -- common/autotest_common.sh@958 -- # kill -0 71116 00:07:56.678 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 958: kill: (71116) - No such process 00:07:56.678 15:13:46 event.cpu_locks -- common/autotest_common.sh@981 -- # echo 'Process with pid 71116 is not found' 00:07:56.678 15:13:46 event.cpu_locks -- event/cpu_locks.sh@18 -- # rm -f 00:07:56.678 00:07:56.678 real 0m21.437s 00:07:56.678 user 0m34.406s 00:07:56.678 sys 0m7.111s 00:07:56.678 ************************************ 00:07:56.678 END TEST cpu_locks 00:07:56.678 ************************************ 00:07:56.678 15:13:46 event.cpu_locks -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:56.678 15:13:46 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:07:56.678 ************************************ 00:07:56.678 END TEST event 00:07:56.678 ************************************ 00:07:56.678 00:07:56.678 real 0m50.035s 00:07:56.678 user 1m33.875s 00:07:56.678 sys 0m11.310s 00:07:56.678 15:13:46 event -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:56.678 15:13:46 event -- common/autotest_common.sh@10 -- # set +x 00:07:56.678 15:13:46 -- spdk/autotest.sh@169 -- # run_test thread /home/vagrant/spdk_repo/spdk/test/thread/thread.sh 00:07:56.678 15:13:46 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:07:56.678 15:13:46 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:56.678 15:13:46 -- common/autotest_common.sh@10 -- # set +x 00:07:56.678 ************************************ 00:07:56.678 START TEST thread 00:07:56.678 ************************************ 00:07:56.678 15:13:46 thread -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/thread/thread.sh 00:07:56.939 * Looking for test storage... 00:07:56.939 * Found test storage at /home/vagrant/spdk_repo/spdk/test/thread 00:07:56.939 15:13:47 thread -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:07:56.939 15:13:47 thread -- common/autotest_common.sh@1693 -- # lcov --version 00:07:56.939 15:13:47 thread -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:07:56.939 15:13:47 thread -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:07:56.939 15:13:47 thread -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:07:56.939 15:13:47 thread -- scripts/common.sh@333 -- # local ver1 ver1_l 00:07:56.939 15:13:47 thread -- scripts/common.sh@334 -- # local ver2 ver2_l 00:07:56.939 15:13:47 thread -- scripts/common.sh@336 -- # IFS=.-: 00:07:56.939 15:13:47 thread -- scripts/common.sh@336 -- # read -ra ver1 00:07:56.939 15:13:47 thread -- scripts/common.sh@337 -- # IFS=.-: 00:07:56.939 15:13:47 thread -- scripts/common.sh@337 -- # read -ra ver2 00:07:56.939 15:13:47 thread -- scripts/common.sh@338 -- # local 'op=<' 00:07:56.939 15:13:47 thread -- scripts/common.sh@340 -- # ver1_l=2 00:07:56.939 15:13:47 thread -- scripts/common.sh@341 -- # ver2_l=1 00:07:56.939 15:13:47 thread -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:07:56.939 15:13:47 thread -- scripts/common.sh@344 -- # case "$op" in 00:07:56.939 15:13:47 thread -- scripts/common.sh@345 -- # : 1 00:07:56.939 15:13:47 thread -- scripts/common.sh@364 -- # (( v = 0 )) 00:07:56.939 15:13:47 thread -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:07:56.939 15:13:47 thread -- scripts/common.sh@365 -- # decimal 1 00:07:56.939 15:13:47 thread -- scripts/common.sh@353 -- # local d=1 00:07:56.939 15:13:47 thread -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:07:56.939 15:13:47 thread -- scripts/common.sh@355 -- # echo 1 00:07:56.939 15:13:47 thread -- scripts/common.sh@365 -- # ver1[v]=1 00:07:56.939 15:13:47 thread -- scripts/common.sh@366 -- # decimal 2 00:07:56.939 15:13:47 thread -- scripts/common.sh@353 -- # local d=2 00:07:56.939 15:13:47 thread -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:07:56.939 15:13:47 thread -- scripts/common.sh@355 -- # echo 2 00:07:56.939 15:13:47 thread -- scripts/common.sh@366 -- # ver2[v]=2 00:07:56.939 15:13:47 thread -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:07:56.939 15:13:47 thread -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:07:56.939 15:13:47 thread -- scripts/common.sh@368 -- # return 0 00:07:56.939 15:13:47 thread -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:07:56.939 15:13:47 thread -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:07:56.939 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:56.939 --rc genhtml_branch_coverage=1 00:07:56.939 --rc genhtml_function_coverage=1 00:07:56.939 --rc genhtml_legend=1 00:07:56.939 --rc geninfo_all_blocks=1 00:07:56.939 --rc geninfo_unexecuted_blocks=1 00:07:56.939 00:07:56.939 ' 00:07:56.939 15:13:47 thread -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:07:56.939 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:56.939 --rc genhtml_branch_coverage=1 00:07:56.939 --rc genhtml_function_coverage=1 00:07:56.939 --rc genhtml_legend=1 00:07:56.939 --rc geninfo_all_blocks=1 00:07:56.939 --rc geninfo_unexecuted_blocks=1 00:07:56.939 00:07:56.939 ' 00:07:56.939 15:13:47 thread -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:07:56.939 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:56.939 --rc genhtml_branch_coverage=1 00:07:56.939 --rc genhtml_function_coverage=1 00:07:56.939 --rc genhtml_legend=1 00:07:56.939 --rc geninfo_all_blocks=1 00:07:56.939 --rc geninfo_unexecuted_blocks=1 00:07:56.939 00:07:56.939 ' 00:07:56.939 15:13:47 thread -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:07:56.939 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:56.939 --rc genhtml_branch_coverage=1 00:07:56.939 --rc genhtml_function_coverage=1 00:07:56.939 --rc genhtml_legend=1 00:07:56.939 --rc geninfo_all_blocks=1 00:07:56.939 --rc geninfo_unexecuted_blocks=1 00:07:56.939 00:07:56.939 ' 00:07:56.939 15:13:47 thread -- thread/thread.sh@11 -- # run_test thread_poller_perf /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 1 -t 1 00:07:56.939 15:13:47 thread -- common/autotest_common.sh@1105 -- # '[' 8 -le 1 ']' 00:07:56.939 15:13:47 thread -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:56.939 15:13:47 thread -- common/autotest_common.sh@10 -- # set +x 00:07:56.939 ************************************ 00:07:56.939 START TEST thread_poller_perf 00:07:56.939 ************************************ 00:07:56.939 15:13:47 thread.thread_poller_perf -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 1 -t 1 00:07:56.939 [2024-11-19 15:13:47.230275] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:07:56.939 [2024-11-19 15:13:47.230489] [ DPDK EAL parameters: poller_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71254 ] 00:07:57.199 [2024-11-19 15:13:47.386536] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:57.199 [2024-11-19 15:13:47.426927] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:57.199 Running 1000 pollers for 1 seconds with 1 microseconds period. 00:07:58.580 [2024-11-19T15:13:48.919Z] ====================================== 00:07:58.580 [2024-11-19T15:13:48.919Z] busy:2299389626 (cyc) 00:07:58.580 [2024-11-19T15:13:48.919Z] total_run_count: 392000 00:07:58.580 [2024-11-19T15:13:48.919Z] tsc_hz: 2290000000 (cyc) 00:07:58.580 [2024-11-19T15:13:48.919Z] ====================================== 00:07:58.580 [2024-11-19T15:13:48.919Z] poller_cost: 5865 (cyc), 2561 (nsec) 00:07:58.580 00:07:58.580 real 0m1.322s 00:07:58.580 user 0m1.140s 00:07:58.580 sys 0m0.076s 00:07:58.580 15:13:48 thread.thread_poller_perf -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:58.580 15:13:48 thread.thread_poller_perf -- common/autotest_common.sh@10 -- # set +x 00:07:58.580 ************************************ 00:07:58.580 END TEST thread_poller_perf 00:07:58.580 ************************************ 00:07:58.580 15:13:48 thread -- thread/thread.sh@12 -- # run_test thread_poller_perf /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 0 -t 1 00:07:58.580 15:13:48 thread -- common/autotest_common.sh@1105 -- # '[' 8 -le 1 ']' 00:07:58.580 15:13:48 thread -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:58.580 15:13:48 thread -- common/autotest_common.sh@10 -- # set +x 00:07:58.580 ************************************ 00:07:58.580 START TEST thread_poller_perf 00:07:58.580 ************************************ 00:07:58.581 15:13:48 thread.thread_poller_perf -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 0 -t 1 00:07:58.581 [2024-11-19 15:13:48.634186] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:07:58.581 [2024-11-19 15:13:48.634391] [ DPDK EAL parameters: poller_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71291 ] 00:07:58.581 [2024-11-19 15:13:48.789723] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:58.581 [2024-11-19 15:13:48.829005] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:58.581 Running 1000 pollers for 1 seconds with 0 microseconds period. 00:07:59.961 [2024-11-19T15:13:50.300Z] ====================================== 00:07:59.961 [2024-11-19T15:13:50.300Z] busy:2293597956 (cyc) 00:07:59.961 [2024-11-19T15:13:50.300Z] total_run_count: 5150000 00:07:59.961 [2024-11-19T15:13:50.300Z] tsc_hz: 2290000000 (cyc) 00:07:59.961 [2024-11-19T15:13:50.300Z] ====================================== 00:07:59.961 [2024-11-19T15:13:50.300Z] poller_cost: 445 (cyc), 194 (nsec) 00:07:59.961 00:07:59.961 real 0m1.321s 00:07:59.961 user 0m1.122s 00:07:59.961 sys 0m0.092s 00:07:59.961 ************************************ 00:07:59.961 END TEST thread_poller_perf 00:07:59.961 ************************************ 00:07:59.961 15:13:49 thread.thread_poller_perf -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:59.961 15:13:49 thread.thread_poller_perf -- common/autotest_common.sh@10 -- # set +x 00:07:59.961 15:13:49 thread -- thread/thread.sh@17 -- # [[ y != \y ]] 00:07:59.961 ************************************ 00:07:59.961 END TEST thread 00:07:59.961 ************************************ 00:07:59.961 00:07:59.961 real 0m3.027s 00:07:59.961 user 0m2.432s 00:07:59.961 sys 0m0.394s 00:07:59.961 15:13:49 thread -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:59.961 15:13:49 thread -- common/autotest_common.sh@10 -- # set +x 00:07:59.961 15:13:50 -- spdk/autotest.sh@171 -- # [[ 0 -eq 1 ]] 00:07:59.961 15:13:50 -- spdk/autotest.sh@176 -- # run_test app_cmdline /home/vagrant/spdk_repo/spdk/test/app/cmdline.sh 00:07:59.961 15:13:50 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:07:59.961 15:13:50 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:59.961 15:13:50 -- common/autotest_common.sh@10 -- # set +x 00:07:59.961 ************************************ 00:07:59.961 START TEST app_cmdline 00:07:59.961 ************************************ 00:07:59.961 15:13:50 app_cmdline -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/app/cmdline.sh 00:07:59.961 * Looking for test storage... 00:07:59.961 * Found test storage at /home/vagrant/spdk_repo/spdk/test/app 00:07:59.961 15:13:50 app_cmdline -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:07:59.961 15:13:50 app_cmdline -- common/autotest_common.sh@1693 -- # lcov --version 00:07:59.961 15:13:50 app_cmdline -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:07:59.961 15:13:50 app_cmdline -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:07:59.961 15:13:50 app_cmdline -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:07:59.961 15:13:50 app_cmdline -- scripts/common.sh@333 -- # local ver1 ver1_l 00:07:59.961 15:13:50 app_cmdline -- scripts/common.sh@334 -- # local ver2 ver2_l 00:07:59.961 15:13:50 app_cmdline -- scripts/common.sh@336 -- # IFS=.-: 00:07:59.961 15:13:50 app_cmdline -- scripts/common.sh@336 -- # read -ra ver1 00:07:59.961 15:13:50 app_cmdline -- scripts/common.sh@337 -- # IFS=.-: 00:07:59.961 15:13:50 app_cmdline -- scripts/common.sh@337 -- # read -ra ver2 00:07:59.961 15:13:50 app_cmdline -- scripts/common.sh@338 -- # local 'op=<' 00:07:59.961 15:13:50 app_cmdline -- scripts/common.sh@340 -- # ver1_l=2 00:07:59.961 15:13:50 app_cmdline -- scripts/common.sh@341 -- # ver2_l=1 00:07:59.961 15:13:50 app_cmdline -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:07:59.961 15:13:50 app_cmdline -- scripts/common.sh@344 -- # case "$op" in 00:07:59.961 15:13:50 app_cmdline -- scripts/common.sh@345 -- # : 1 00:07:59.961 15:13:50 app_cmdline -- scripts/common.sh@364 -- # (( v = 0 )) 00:07:59.961 15:13:50 app_cmdline -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:07:59.961 15:13:50 app_cmdline -- scripts/common.sh@365 -- # decimal 1 00:07:59.961 15:13:50 app_cmdline -- scripts/common.sh@353 -- # local d=1 00:07:59.961 15:13:50 app_cmdline -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:07:59.961 15:13:50 app_cmdline -- scripts/common.sh@355 -- # echo 1 00:07:59.961 15:13:50 app_cmdline -- scripts/common.sh@365 -- # ver1[v]=1 00:07:59.961 15:13:50 app_cmdline -- scripts/common.sh@366 -- # decimal 2 00:07:59.961 15:13:50 app_cmdline -- scripts/common.sh@353 -- # local d=2 00:07:59.961 15:13:50 app_cmdline -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:07:59.961 15:13:50 app_cmdline -- scripts/common.sh@355 -- # echo 2 00:07:59.961 15:13:50 app_cmdline -- scripts/common.sh@366 -- # ver2[v]=2 00:07:59.961 15:13:50 app_cmdline -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:07:59.961 15:13:50 app_cmdline -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:07:59.961 15:13:50 app_cmdline -- scripts/common.sh@368 -- # return 0 00:07:59.961 15:13:50 app_cmdline -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:07:59.961 15:13:50 app_cmdline -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:07:59.961 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:59.961 --rc genhtml_branch_coverage=1 00:07:59.961 --rc genhtml_function_coverage=1 00:07:59.961 --rc genhtml_legend=1 00:07:59.961 --rc geninfo_all_blocks=1 00:07:59.961 --rc geninfo_unexecuted_blocks=1 00:07:59.961 00:07:59.961 ' 00:07:59.961 15:13:50 app_cmdline -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:07:59.961 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:59.961 --rc genhtml_branch_coverage=1 00:07:59.961 --rc genhtml_function_coverage=1 00:07:59.961 --rc genhtml_legend=1 00:07:59.961 --rc geninfo_all_blocks=1 00:07:59.961 --rc geninfo_unexecuted_blocks=1 00:07:59.961 00:07:59.961 ' 00:07:59.961 15:13:50 app_cmdline -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:07:59.961 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:59.961 --rc genhtml_branch_coverage=1 00:07:59.961 --rc genhtml_function_coverage=1 00:07:59.961 --rc genhtml_legend=1 00:07:59.961 --rc geninfo_all_blocks=1 00:07:59.961 --rc geninfo_unexecuted_blocks=1 00:07:59.961 00:07:59.961 ' 00:07:59.961 15:13:50 app_cmdline -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:07:59.961 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:59.961 --rc genhtml_branch_coverage=1 00:07:59.961 --rc genhtml_function_coverage=1 00:07:59.961 --rc genhtml_legend=1 00:07:59.961 --rc geninfo_all_blocks=1 00:07:59.961 --rc geninfo_unexecuted_blocks=1 00:07:59.961 00:07:59.961 ' 00:07:59.961 15:13:50 app_cmdline -- app/cmdline.sh@14 -- # trap 'killprocess $spdk_tgt_pid' EXIT 00:07:59.962 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:59.962 15:13:50 app_cmdline -- app/cmdline.sh@17 -- # spdk_tgt_pid=71374 00:07:59.962 15:13:50 app_cmdline -- app/cmdline.sh@16 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --rpcs-allowed spdk_get_version,rpc_get_methods 00:07:59.962 15:13:50 app_cmdline -- app/cmdline.sh@18 -- # waitforlisten 71374 00:07:59.962 15:13:50 app_cmdline -- common/autotest_common.sh@835 -- # '[' -z 71374 ']' 00:07:59.962 15:13:50 app_cmdline -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:59.962 15:13:50 app_cmdline -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:59.962 15:13:50 app_cmdline -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:59.962 15:13:50 app_cmdline -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:59.962 15:13:50 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:08:00.222 [2024-11-19 15:13:50.374170] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:08:00.222 [2024-11-19 15:13:50.374423] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71374 ] 00:08:00.222 [2024-11-19 15:13:50.531635] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:00.482 [2024-11-19 15:13:50.571758] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:01.051 15:13:51 app_cmdline -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:01.051 15:13:51 app_cmdline -- common/autotest_common.sh@868 -- # return 0 00:08:01.051 15:13:51 app_cmdline -- app/cmdline.sh@20 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py spdk_get_version 00:08:01.310 { 00:08:01.310 "version": "SPDK v25.01-pre git sha1 dcc2ca8f3", 00:08:01.310 "fields": { 00:08:01.310 "major": 25, 00:08:01.310 "minor": 1, 00:08:01.310 "patch": 0, 00:08:01.310 "suffix": "-pre", 00:08:01.310 "commit": "dcc2ca8f3" 00:08:01.310 } 00:08:01.310 } 00:08:01.310 15:13:51 app_cmdline -- app/cmdline.sh@22 -- # expected_methods=() 00:08:01.310 15:13:51 app_cmdline -- app/cmdline.sh@23 -- # expected_methods+=("rpc_get_methods") 00:08:01.310 15:13:51 app_cmdline -- app/cmdline.sh@24 -- # expected_methods+=("spdk_get_version") 00:08:01.310 15:13:51 app_cmdline -- app/cmdline.sh@26 -- # methods=($(rpc_cmd rpc_get_methods | jq -r ".[]" | sort)) 00:08:01.310 15:13:51 app_cmdline -- app/cmdline.sh@26 -- # rpc_cmd rpc_get_methods 00:08:01.310 15:13:51 app_cmdline -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:01.310 15:13:51 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:08:01.310 15:13:51 app_cmdline -- app/cmdline.sh@26 -- # jq -r '.[]' 00:08:01.310 15:13:51 app_cmdline -- app/cmdline.sh@26 -- # sort 00:08:01.310 15:13:51 app_cmdline -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:01.310 15:13:51 app_cmdline -- app/cmdline.sh@27 -- # (( 2 == 2 )) 00:08:01.310 15:13:51 app_cmdline -- app/cmdline.sh@28 -- # [[ rpc_get_methods spdk_get_version == \r\p\c\_\g\e\t\_\m\e\t\h\o\d\s\ \s\p\d\k\_\g\e\t\_\v\e\r\s\i\o\n ]] 00:08:01.310 15:13:51 app_cmdline -- app/cmdline.sh@30 -- # NOT /home/vagrant/spdk_repo/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:08:01.310 15:13:51 app_cmdline -- common/autotest_common.sh@652 -- # local es=0 00:08:01.310 15:13:51 app_cmdline -- common/autotest_common.sh@654 -- # valid_exec_arg /home/vagrant/spdk_repo/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:08:01.310 15:13:51 app_cmdline -- common/autotest_common.sh@640 -- # local arg=/home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:08:01.310 15:13:51 app_cmdline -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:08:01.310 15:13:51 app_cmdline -- common/autotest_common.sh@644 -- # type -t /home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:08:01.310 15:13:51 app_cmdline -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:08:01.310 15:13:51 app_cmdline -- common/autotest_common.sh@646 -- # type -P /home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:08:01.310 15:13:51 app_cmdline -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:08:01.310 15:13:51 app_cmdline -- common/autotest_common.sh@646 -- # arg=/home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:08:01.310 15:13:51 app_cmdline -- common/autotest_common.sh@646 -- # [[ -x /home/vagrant/spdk_repo/spdk/scripts/rpc.py ]] 00:08:01.310 15:13:51 app_cmdline -- common/autotest_common.sh@655 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:08:01.577 request: 00:08:01.577 { 00:08:01.577 "method": "env_dpdk_get_mem_stats", 00:08:01.577 "req_id": 1 00:08:01.577 } 00:08:01.577 Got JSON-RPC error response 00:08:01.577 response: 00:08:01.577 { 00:08:01.577 "code": -32601, 00:08:01.577 "message": "Method not found" 00:08:01.577 } 00:08:01.577 15:13:51 app_cmdline -- common/autotest_common.sh@655 -- # es=1 00:08:01.577 15:13:51 app_cmdline -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:08:01.577 15:13:51 app_cmdline -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:08:01.577 15:13:51 app_cmdline -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:08:01.577 15:13:51 app_cmdline -- app/cmdline.sh@1 -- # killprocess 71374 00:08:01.577 15:13:51 app_cmdline -- common/autotest_common.sh@954 -- # '[' -z 71374 ']' 00:08:01.577 15:13:51 app_cmdline -- common/autotest_common.sh@958 -- # kill -0 71374 00:08:01.577 15:13:51 app_cmdline -- common/autotest_common.sh@959 -- # uname 00:08:01.577 15:13:51 app_cmdline -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:01.577 15:13:51 app_cmdline -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71374 00:08:01.577 15:13:51 app_cmdline -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:01.577 15:13:51 app_cmdline -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:01.577 15:13:51 app_cmdline -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71374' 00:08:01.577 killing process with pid 71374 00:08:01.577 15:13:51 app_cmdline -- common/autotest_common.sh@973 -- # kill 71374 00:08:01.577 15:13:51 app_cmdline -- common/autotest_common.sh@978 -- # wait 71374 00:08:02.155 00:08:02.155 real 0m2.290s 00:08:02.155 user 0m2.412s 00:08:02.155 sys 0m0.697s 00:08:02.155 15:13:52 app_cmdline -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:02.156 15:13:52 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:08:02.156 ************************************ 00:08:02.156 END TEST app_cmdline 00:08:02.156 ************************************ 00:08:02.156 15:13:52 -- spdk/autotest.sh@177 -- # run_test version /home/vagrant/spdk_repo/spdk/test/app/version.sh 00:08:02.156 15:13:52 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:08:02.156 15:13:52 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:02.156 15:13:52 -- common/autotest_common.sh@10 -- # set +x 00:08:02.156 ************************************ 00:08:02.156 START TEST version 00:08:02.156 ************************************ 00:08:02.156 15:13:52 version -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/app/version.sh 00:08:02.414 * Looking for test storage... 00:08:02.414 * Found test storage at /home/vagrant/spdk_repo/spdk/test/app 00:08:02.414 15:13:52 version -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:08:02.414 15:13:52 version -- common/autotest_common.sh@1693 -- # lcov --version 00:08:02.414 15:13:52 version -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:08:02.414 15:13:52 version -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:08:02.414 15:13:52 version -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:08:02.414 15:13:52 version -- scripts/common.sh@333 -- # local ver1 ver1_l 00:08:02.414 15:13:52 version -- scripts/common.sh@334 -- # local ver2 ver2_l 00:08:02.414 15:13:52 version -- scripts/common.sh@336 -- # IFS=.-: 00:08:02.414 15:13:52 version -- scripts/common.sh@336 -- # read -ra ver1 00:08:02.414 15:13:52 version -- scripts/common.sh@337 -- # IFS=.-: 00:08:02.414 15:13:52 version -- scripts/common.sh@337 -- # read -ra ver2 00:08:02.414 15:13:52 version -- scripts/common.sh@338 -- # local 'op=<' 00:08:02.414 15:13:52 version -- scripts/common.sh@340 -- # ver1_l=2 00:08:02.414 15:13:52 version -- scripts/common.sh@341 -- # ver2_l=1 00:08:02.414 15:13:52 version -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:08:02.414 15:13:52 version -- scripts/common.sh@344 -- # case "$op" in 00:08:02.414 15:13:52 version -- scripts/common.sh@345 -- # : 1 00:08:02.414 15:13:52 version -- scripts/common.sh@364 -- # (( v = 0 )) 00:08:02.414 15:13:52 version -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:08:02.414 15:13:52 version -- scripts/common.sh@365 -- # decimal 1 00:08:02.414 15:13:52 version -- scripts/common.sh@353 -- # local d=1 00:08:02.414 15:13:52 version -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:08:02.414 15:13:52 version -- scripts/common.sh@355 -- # echo 1 00:08:02.414 15:13:52 version -- scripts/common.sh@365 -- # ver1[v]=1 00:08:02.414 15:13:52 version -- scripts/common.sh@366 -- # decimal 2 00:08:02.414 15:13:52 version -- scripts/common.sh@353 -- # local d=2 00:08:02.414 15:13:52 version -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:08:02.414 15:13:52 version -- scripts/common.sh@355 -- # echo 2 00:08:02.414 15:13:52 version -- scripts/common.sh@366 -- # ver2[v]=2 00:08:02.415 15:13:52 version -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:08:02.415 15:13:52 version -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:08:02.415 15:13:52 version -- scripts/common.sh@368 -- # return 0 00:08:02.415 15:13:52 version -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:08:02.415 15:13:52 version -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:08:02.415 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:02.415 --rc genhtml_branch_coverage=1 00:08:02.415 --rc genhtml_function_coverage=1 00:08:02.415 --rc genhtml_legend=1 00:08:02.415 --rc geninfo_all_blocks=1 00:08:02.415 --rc geninfo_unexecuted_blocks=1 00:08:02.415 00:08:02.415 ' 00:08:02.415 15:13:52 version -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:08:02.415 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:02.415 --rc genhtml_branch_coverage=1 00:08:02.415 --rc genhtml_function_coverage=1 00:08:02.415 --rc genhtml_legend=1 00:08:02.415 --rc geninfo_all_blocks=1 00:08:02.415 --rc geninfo_unexecuted_blocks=1 00:08:02.415 00:08:02.415 ' 00:08:02.415 15:13:52 version -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:08:02.415 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:02.415 --rc genhtml_branch_coverage=1 00:08:02.415 --rc genhtml_function_coverage=1 00:08:02.415 --rc genhtml_legend=1 00:08:02.415 --rc geninfo_all_blocks=1 00:08:02.415 --rc geninfo_unexecuted_blocks=1 00:08:02.415 00:08:02.415 ' 00:08:02.415 15:13:52 version -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:08:02.415 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:02.415 --rc genhtml_branch_coverage=1 00:08:02.415 --rc genhtml_function_coverage=1 00:08:02.415 --rc genhtml_legend=1 00:08:02.415 --rc geninfo_all_blocks=1 00:08:02.415 --rc geninfo_unexecuted_blocks=1 00:08:02.415 00:08:02.415 ' 00:08:02.415 15:13:52 version -- app/version.sh@17 -- # get_header_version major 00:08:02.415 15:13:52 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_MAJOR[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:08:02.415 15:13:52 version -- app/version.sh@14 -- # cut -f2 00:08:02.415 15:13:52 version -- app/version.sh@14 -- # tr -d '"' 00:08:02.415 15:13:52 version -- app/version.sh@17 -- # major=25 00:08:02.415 15:13:52 version -- app/version.sh@18 -- # get_header_version minor 00:08:02.415 15:13:52 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_MINOR[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:08:02.415 15:13:52 version -- app/version.sh@14 -- # cut -f2 00:08:02.415 15:13:52 version -- app/version.sh@14 -- # tr -d '"' 00:08:02.415 15:13:52 version -- app/version.sh@18 -- # minor=1 00:08:02.415 15:13:52 version -- app/version.sh@19 -- # get_header_version patch 00:08:02.415 15:13:52 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_PATCH[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:08:02.415 15:13:52 version -- app/version.sh@14 -- # cut -f2 00:08:02.415 15:13:52 version -- app/version.sh@14 -- # tr -d '"' 00:08:02.415 15:13:52 version -- app/version.sh@19 -- # patch=0 00:08:02.415 15:13:52 version -- app/version.sh@20 -- # get_header_version suffix 00:08:02.415 15:13:52 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_SUFFIX[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:08:02.415 15:13:52 version -- app/version.sh@14 -- # cut -f2 00:08:02.415 15:13:52 version -- app/version.sh@14 -- # tr -d '"' 00:08:02.415 15:13:52 version -- app/version.sh@20 -- # suffix=-pre 00:08:02.415 15:13:52 version -- app/version.sh@22 -- # version=25.1 00:08:02.415 15:13:52 version -- app/version.sh@25 -- # (( patch != 0 )) 00:08:02.415 15:13:52 version -- app/version.sh@28 -- # version=25.1rc0 00:08:02.415 15:13:52 version -- app/version.sh@30 -- # PYTHONPATH=:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python 00:08:02.415 15:13:52 version -- app/version.sh@30 -- # python3 -c 'import spdk; print(spdk.__version__)' 00:08:02.415 15:13:52 version -- app/version.sh@30 -- # py_version=25.1rc0 00:08:02.415 15:13:52 version -- app/version.sh@31 -- # [[ 25.1rc0 == \2\5\.\1\r\c\0 ]] 00:08:02.415 ************************************ 00:08:02.415 END TEST version 00:08:02.415 ************************************ 00:08:02.415 00:08:02.415 real 0m0.310s 00:08:02.415 user 0m0.178s 00:08:02.415 sys 0m0.188s 00:08:02.415 15:13:52 version -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:02.415 15:13:52 version -- common/autotest_common.sh@10 -- # set +x 00:08:02.674 15:13:52 -- spdk/autotest.sh@179 -- # '[' 0 -eq 1 ']' 00:08:02.674 15:13:52 -- spdk/autotest.sh@188 -- # [[ 1 -eq 1 ]] 00:08:02.674 15:13:52 -- spdk/autotest.sh@189 -- # run_test bdev_raid /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh 00:08:02.674 15:13:52 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:08:02.674 15:13:52 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:02.674 15:13:52 -- common/autotest_common.sh@10 -- # set +x 00:08:02.674 ************************************ 00:08:02.674 START TEST bdev_raid 00:08:02.674 ************************************ 00:08:02.675 15:13:52 bdev_raid -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh 00:08:02.675 * Looking for test storage... 00:08:02.675 * Found test storage at /home/vagrant/spdk_repo/spdk/test/bdev 00:08:02.675 15:13:52 bdev_raid -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:08:02.675 15:13:52 bdev_raid -- common/autotest_common.sh@1693 -- # lcov --version 00:08:02.675 15:13:52 bdev_raid -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:08:02.675 15:13:52 bdev_raid -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:08:02.675 15:13:52 bdev_raid -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:08:02.675 15:13:52 bdev_raid -- scripts/common.sh@333 -- # local ver1 ver1_l 00:08:02.675 15:13:52 bdev_raid -- scripts/common.sh@334 -- # local ver2 ver2_l 00:08:02.675 15:13:52 bdev_raid -- scripts/common.sh@336 -- # IFS=.-: 00:08:02.675 15:13:52 bdev_raid -- scripts/common.sh@336 -- # read -ra ver1 00:08:02.675 15:13:52 bdev_raid -- scripts/common.sh@337 -- # IFS=.-: 00:08:02.675 15:13:52 bdev_raid -- scripts/common.sh@337 -- # read -ra ver2 00:08:02.675 15:13:52 bdev_raid -- scripts/common.sh@338 -- # local 'op=<' 00:08:02.675 15:13:52 bdev_raid -- scripts/common.sh@340 -- # ver1_l=2 00:08:02.675 15:13:52 bdev_raid -- scripts/common.sh@341 -- # ver2_l=1 00:08:02.675 15:13:52 bdev_raid -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:08:02.675 15:13:52 bdev_raid -- scripts/common.sh@344 -- # case "$op" in 00:08:02.675 15:13:52 bdev_raid -- scripts/common.sh@345 -- # : 1 00:08:02.675 15:13:52 bdev_raid -- scripts/common.sh@364 -- # (( v = 0 )) 00:08:02.675 15:13:52 bdev_raid -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:08:02.675 15:13:52 bdev_raid -- scripts/common.sh@365 -- # decimal 1 00:08:02.675 15:13:52 bdev_raid -- scripts/common.sh@353 -- # local d=1 00:08:02.675 15:13:52 bdev_raid -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:08:02.675 15:13:52 bdev_raid -- scripts/common.sh@355 -- # echo 1 00:08:02.675 15:13:52 bdev_raid -- scripts/common.sh@365 -- # ver1[v]=1 00:08:02.675 15:13:52 bdev_raid -- scripts/common.sh@366 -- # decimal 2 00:08:02.675 15:13:52 bdev_raid -- scripts/common.sh@353 -- # local d=2 00:08:02.675 15:13:52 bdev_raid -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:08:02.675 15:13:52 bdev_raid -- scripts/common.sh@355 -- # echo 2 00:08:02.675 15:13:53 bdev_raid -- scripts/common.sh@366 -- # ver2[v]=2 00:08:02.675 15:13:53 bdev_raid -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:08:02.675 15:13:53 bdev_raid -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:08:02.675 15:13:53 bdev_raid -- scripts/common.sh@368 -- # return 0 00:08:02.675 15:13:53 bdev_raid -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:08:02.675 15:13:53 bdev_raid -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:08:02.675 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:02.675 --rc genhtml_branch_coverage=1 00:08:02.675 --rc genhtml_function_coverage=1 00:08:02.675 --rc genhtml_legend=1 00:08:02.675 --rc geninfo_all_blocks=1 00:08:02.675 --rc geninfo_unexecuted_blocks=1 00:08:02.675 00:08:02.675 ' 00:08:02.675 15:13:53 bdev_raid -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:08:02.675 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:02.675 --rc genhtml_branch_coverage=1 00:08:02.675 --rc genhtml_function_coverage=1 00:08:02.675 --rc genhtml_legend=1 00:08:02.675 --rc geninfo_all_blocks=1 00:08:02.675 --rc geninfo_unexecuted_blocks=1 00:08:02.675 00:08:02.675 ' 00:08:02.675 15:13:53 bdev_raid -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:08:02.675 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:02.675 --rc genhtml_branch_coverage=1 00:08:02.675 --rc genhtml_function_coverage=1 00:08:02.675 --rc genhtml_legend=1 00:08:02.675 --rc geninfo_all_blocks=1 00:08:02.675 --rc geninfo_unexecuted_blocks=1 00:08:02.675 00:08:02.675 ' 00:08:02.675 15:13:53 bdev_raid -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:08:02.675 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:02.675 --rc genhtml_branch_coverage=1 00:08:02.675 --rc genhtml_function_coverage=1 00:08:02.675 --rc genhtml_legend=1 00:08:02.675 --rc geninfo_all_blocks=1 00:08:02.675 --rc geninfo_unexecuted_blocks=1 00:08:02.675 00:08:02.675 ' 00:08:02.675 15:13:53 bdev_raid -- bdev/bdev_raid.sh@12 -- # source /home/vagrant/spdk_repo/spdk/test/bdev/nbd_common.sh 00:08:02.675 15:13:53 bdev_raid -- bdev/nbd_common.sh@6 -- # set -e 00:08:02.675 15:13:53 bdev_raid -- bdev/bdev_raid.sh@14 -- # rpc_py=rpc_cmd 00:08:02.935 15:13:53 bdev_raid -- bdev/bdev_raid.sh@946 -- # mkdir -p /raidtest 00:08:02.935 15:13:53 bdev_raid -- bdev/bdev_raid.sh@947 -- # trap 'cleanup; exit 1' EXIT 00:08:02.935 15:13:53 bdev_raid -- bdev/bdev_raid.sh@949 -- # base_blocklen=512 00:08:02.935 15:13:53 bdev_raid -- bdev/bdev_raid.sh@951 -- # run_test raid1_resize_data_offset_test raid_resize_data_offset_test 00:08:02.935 15:13:53 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:08:02.935 15:13:53 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:02.935 15:13:53 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:02.935 ************************************ 00:08:02.935 START TEST raid1_resize_data_offset_test 00:08:02.935 ************************************ 00:08:02.935 Process raid pid: 71545 00:08:02.935 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:02.935 15:13:53 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@1129 -- # raid_resize_data_offset_test 00:08:02.935 15:13:53 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@917 -- # raid_pid=71545 00:08:02.935 15:13:53 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@918 -- # echo 'Process raid pid: 71545' 00:08:02.935 15:13:53 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@919 -- # waitforlisten 71545 00:08:02.935 15:13:53 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@835 -- # '[' -z 71545 ']' 00:08:02.935 15:13:53 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@916 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:02.935 15:13:53 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:02.935 15:13:53 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:02.935 15:13:53 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:02.935 15:13:53 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:02.935 15:13:53 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.935 [2024-11-19 15:13:53.114124] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:08:02.935 [2024-11-19 15:13:53.114350] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:02.935 [2024-11-19 15:13:53.272890] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:03.196 [2024-11-19 15:13:53.317348] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:03.196 [2024-11-19 15:13:53.393433] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:03.196 [2024-11-19 15:13:53.393591] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:03.764 15:13:53 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:03.764 15:13:53 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@868 -- # return 0 00:08:03.764 15:13:53 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@922 -- # rpc_cmd bdev_malloc_create -b malloc0 64 512 -o 16 00:08:03.764 15:13:53 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:03.764 15:13:53 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:08:03.764 malloc0 00:08:03.764 15:13:53 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:03.764 15:13:53 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@923 -- # rpc_cmd bdev_malloc_create -b malloc1 64 512 -o 16 00:08:03.764 15:13:53 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:03.764 15:13:53 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:08:03.764 malloc1 00:08:03.764 15:13:53 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:03.764 15:13:53 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@924 -- # rpc_cmd bdev_null_create null0 64 512 00:08:03.764 15:13:53 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:03.764 15:13:53 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:08:03.764 null0 00:08:03.764 15:13:54 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:03.764 15:13:54 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@926 -- # rpc_cmd bdev_raid_create -n Raid -r 1 -b ''\''malloc0 malloc1 null0'\''' -s 00:08:03.764 15:13:54 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:03.764 15:13:54 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:08:03.764 [2024-11-19 15:13:54.017330] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc0 is claimed 00:08:03.764 [2024-11-19 15:13:54.019693] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:08:03.764 [2024-11-19 15:13:54.019817] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev null0 is claimed 00:08:03.764 [2024-11-19 15:13:54.020050] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:08:03.764 [2024-11-19 15:13:54.020101] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 129024, blocklen 512 00:08:03.764 [2024-11-19 15:13:54.020499] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000021f0 00:08:03.764 [2024-11-19 15:13:54.020702] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:08:03.764 [2024-11-19 15:13:54.020749] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000001200 00:08:03.764 [2024-11-19 15:13:54.021002] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:03.765 15:13:54 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:03.765 15:13:54 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@929 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:03.765 15:13:54 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:03.765 15:13:54 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:08:03.765 15:13:54 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@929 -- # jq -r '.[].base_bdevs_list[2].data_offset' 00:08:03.765 15:13:54 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:03.765 15:13:54 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@929 -- # (( 2048 == 2048 )) 00:08:03.765 15:13:54 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@931 -- # rpc_cmd bdev_null_delete null0 00:08:03.765 15:13:54 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:03.765 15:13:54 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:08:03.765 [2024-11-19 15:13:54.073243] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: null0 00:08:03.765 15:13:54 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:03.765 15:13:54 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@935 -- # rpc_cmd bdev_malloc_create -b malloc2 512 512 -o 30 00:08:03.765 15:13:54 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:03.765 15:13:54 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:08:04.025 malloc2 00:08:04.025 15:13:54 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:04.025 15:13:54 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@936 -- # rpc_cmd bdev_raid_add_base_bdev Raid malloc2 00:08:04.025 15:13:54 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:04.025 15:13:54 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:08:04.025 [2024-11-19 15:13:54.285348] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:08:04.025 [2024-11-19 15:13:54.294688] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:08:04.025 15:13:54 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:04.025 [2024-11-19 15:13:54.297134] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev Raid 00:08:04.025 15:13:54 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@939 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:04.025 15:13:54 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@939 -- # jq -r '.[].base_bdevs_list[2].data_offset' 00:08:04.025 15:13:54 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:04.025 15:13:54 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:08:04.025 15:13:54 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:04.025 15:13:54 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@939 -- # (( 2070 == 2070 )) 00:08:04.025 15:13:54 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@941 -- # killprocess 71545 00:08:04.025 15:13:54 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@954 -- # '[' -z 71545 ']' 00:08:04.025 15:13:54 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@958 -- # kill -0 71545 00:08:04.025 15:13:54 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@959 -- # uname 00:08:04.025 15:13:54 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:04.025 15:13:54 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71545 00:08:04.285 killing process with pid 71545 00:08:04.285 15:13:54 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:04.285 15:13:54 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:04.285 15:13:54 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71545' 00:08:04.285 15:13:54 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@973 -- # kill 71545 00:08:04.285 15:13:54 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@978 -- # wait 71545 00:08:04.286 [2024-11-19 15:13:54.394727] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:04.286 [2024-11-19 15:13:54.396666] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev Raid: Operation canceled 00:08:04.286 [2024-11-19 15:13:54.396750] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:04.286 [2024-11-19 15:13:54.396770] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: malloc2 00:08:04.286 [2024-11-19 15:13:54.407951] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:04.286 [2024-11-19 15:13:54.408322] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:04.286 [2024-11-19 15:13:54.408341] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Raid, state offline 00:08:04.545 [2024-11-19 15:13:54.807575] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:04.805 15:13:55 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@943 -- # return 0 00:08:04.805 00:08:04.805 real 0m2.099s 00:08:04.805 user 0m1.880s 00:08:04.805 sys 0m0.646s 00:08:04.805 15:13:55 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:04.805 ************************************ 00:08:04.805 END TEST raid1_resize_data_offset_test 00:08:04.805 ************************************ 00:08:04.805 15:13:55 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:08:05.064 15:13:55 bdev_raid -- bdev/bdev_raid.sh@953 -- # run_test raid0_resize_superblock_test raid_resize_superblock_test 0 00:08:05.064 15:13:55 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:08:05.064 15:13:55 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:05.064 15:13:55 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:05.064 ************************************ 00:08:05.064 START TEST raid0_resize_superblock_test 00:08:05.064 ************************************ 00:08:05.064 15:13:55 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@1129 -- # raid_resize_superblock_test 0 00:08:05.064 15:13:55 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@854 -- # local raid_level=0 00:08:05.064 15:13:55 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@857 -- # raid_pid=71601 00:08:05.064 Process raid pid: 71601 00:08:05.064 15:13:55 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@856 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:05.064 15:13:55 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@858 -- # echo 'Process raid pid: 71601' 00:08:05.064 15:13:55 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@859 -- # waitforlisten 71601 00:08:05.064 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:05.064 15:13:55 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 71601 ']' 00:08:05.064 15:13:55 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:05.064 15:13:55 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:05.064 15:13:55 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:05.064 15:13:55 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:05.064 15:13:55 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:05.064 [2024-11-19 15:13:55.280754] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:08:05.064 [2024-11-19 15:13:55.280874] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:05.323 [2024-11-19 15:13:55.438495] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:05.323 [2024-11-19 15:13:55.479593] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:05.323 [2024-11-19 15:13:55.555805] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:05.323 [2024-11-19 15:13:55.555857] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:05.892 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:05.892 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:08:05.892 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@861 -- # rpc_cmd bdev_malloc_create -b malloc0 512 512 00:08:05.892 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:05.892 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.153 malloc0 00:08:06.153 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:06.153 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@863 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:08:06.153 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:06.153 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.153 [2024-11-19 15:13:56.336460] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:08:06.153 [2024-11-19 15:13:56.336549] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:06.153 [2024-11-19 15:13:56.336574] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:08:06.153 [2024-11-19 15:13:56.336586] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:06.153 [2024-11-19 15:13:56.339189] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:06.153 [2024-11-19 15:13:56.339233] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:08:06.153 pt0 00:08:06.153 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:06.153 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@864 -- # rpc_cmd bdev_lvol_create_lvstore pt0 lvs0 00:08:06.153 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:06.153 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.413 c5538655-06be-4e2a-89f8-5ffcba889929 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@866 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol0 64 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.413 94e4d481-62dd-483b-8496-8521395bf536 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@867 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol1 64 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.413 119036d8-0c34-4d98-b962-921b90b1698a 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@869 -- # case $raid_level in 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@870 -- # rpc_cmd bdev_raid_create -n Raid -r 0 -z 64 -b ''\''lvs0/lvol0 lvs0/lvol1'\''' -s 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.413 [2024-11-19 15:13:56.551599] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev 94e4d481-62dd-483b-8496-8521395bf536 is claimed 00:08:06.413 [2024-11-19 15:13:56.551834] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev 119036d8-0c34-4d98-b962-921b90b1698a is claimed 00:08:06.413 [2024-11-19 15:13:56.551961] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:08:06.413 [2024-11-19 15:13:56.551995] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 245760, blocklen 512 00:08:06.413 [2024-11-19 15:13:56.552306] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:08:06.413 [2024-11-19 15:13:56.552484] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:08:06.413 [2024-11-19 15:13:56.552496] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000001200 00:08:06.413 [2024-11-19 15:13:56.552663] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # jq '.[].num_blocks' 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # (( 64 == 64 )) 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # jq '.[].num_blocks' 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # (( 64 == 64 )) 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@880 -- # rpc_cmd bdev_get_bdevs -b Raid 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@880 -- # jq '.[].num_blocks' 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.413 [2024-11-19 15:13:56.683574] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@880 -- # (( 245760 == 245760 )) 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@885 -- # rpc_cmd bdev_lvol_resize lvs0/lvol0 100 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.413 [2024-11-19 15:13:56.727431] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:08:06.413 [2024-11-19 15:13:56.727457] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev '94e4d481-62dd-483b-8496-8521395bf536' was resized: old size 131072, new size 204800 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@886 -- # rpc_cmd bdev_lvol_resize lvs0/lvol1 100 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.413 [2024-11-19 15:13:56.739360] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:08:06.413 [2024-11-19 15:13:56.739387] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev '119036d8-0c34-4d98-b962-921b90b1698a' was resized: old size 131072, new size 204800 00:08:06.413 [2024-11-19 15:13:56.739419] bdev_raid.c:2344:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 245760 to 393216 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.413 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # jq '.[].num_blocks' 00:08:06.673 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:06.673 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # (( 100 == 100 )) 00:08:06.673 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # jq '.[].num_blocks' 00:08:06.674 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:08:06.674 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:06.674 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.674 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:06.674 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # (( 100 == 100 )) 00:08:06.674 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:08:06.674 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@894 -- # rpc_cmd bdev_get_bdevs -b Raid 00:08:06.674 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:08:06.674 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@894 -- # jq '.[].num_blocks' 00:08:06.674 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:06.674 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.674 [2024-11-19 15:13:56.851266] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:06.674 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:06.674 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:08:06.674 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:08:06.674 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@894 -- # (( 393216 == 393216 )) 00:08:06.674 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@898 -- # rpc_cmd bdev_passthru_delete pt0 00:08:06.674 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:06.674 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.674 [2024-11-19 15:13:56.898985] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev pt0 being removed: closing lvstore lvs0 00:08:06.674 [2024-11-19 15:13:56.899114] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol0 00:08:06.674 [2024-11-19 15:13:56.899162] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:06.674 [2024-11-19 15:13:56.899199] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol1 00:08:06.674 [2024-11-19 15:13:56.899362] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:06.674 [2024-11-19 15:13:56.899432] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:06.674 [2024-11-19 15:13:56.899482] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Raid, state offline 00:08:06.674 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:06.674 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@899 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:08:06.674 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:06.674 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.674 [2024-11-19 15:13:56.910903] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:08:06.674 [2024-11-19 15:13:56.910973] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:06.674 [2024-11-19 15:13:56.910995] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:08:06.674 [2024-11-19 15:13:56.911007] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:06.674 [2024-11-19 15:13:56.913526] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:06.674 [2024-11-19 15:13:56.913565] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:08:06.674 [2024-11-19 15:13:56.915082] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev 94e4d481-62dd-483b-8496-8521395bf536 00:08:06.674 [2024-11-19 15:13:56.915148] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev 94e4d481-62dd-483b-8496-8521395bf536 is claimed 00:08:06.674 [2024-11-19 15:13:56.915247] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev 119036d8-0c34-4d98-b962-921b90b1698a 00:08:06.674 [2024-11-19 15:13:56.915270] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev 119036d8-0c34-4d98-b962-921b90b1698a is claimed 00:08:06.674 [2024-11-19 15:13:56.915400] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev 119036d8-0c34-4d98-b962-921b90b1698a (2) smaller than existing raid bdev Raid (3) 00:08:06.674 [2024-11-19 15:13:56.915425] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev 94e4d481-62dd-483b-8496-8521395bf536: File exists 00:08:06.674 [2024-11-19 15:13:56.915460] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:08:06.674 [2024-11-19 15:13:56.915470] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 393216, blocklen 512 00:08:06.674 [2024-11-19 15:13:56.915724] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:08:06.674 pt0 00:08:06.674 [2024-11-19 15:13:56.915916] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:08:06.674 [2024-11-19 15:13:56.915927] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000001580 00:08:06.674 [2024-11-19 15:13:56.916064] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:06.674 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:06.674 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@900 -- # rpc_cmd bdev_wait_for_examine 00:08:06.674 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:06.674 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.674 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:06.674 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:08:06.674 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@905 -- # rpc_cmd bdev_get_bdevs -b Raid 00:08:06.674 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:06.674 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.674 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:08:06.674 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@905 -- # jq '.[].num_blocks' 00:08:06.674 [2024-11-19 15:13:56.935323] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:06.674 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:06.674 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:08:06.674 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:08:06.674 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@905 -- # (( 393216 == 393216 )) 00:08:06.674 15:13:56 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@909 -- # killprocess 71601 00:08:06.674 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 71601 ']' 00:08:06.674 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@958 -- # kill -0 71601 00:08:06.674 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@959 -- # uname 00:08:06.674 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:06.674 15:13:56 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71601 00:08:06.934 killing process with pid 71601 00:08:06.934 15:13:57 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:06.934 15:13:57 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:06.934 15:13:57 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71601' 00:08:06.934 15:13:57 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@973 -- # kill 71601 00:08:06.934 [2024-11-19 15:13:57.013521] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:06.934 [2024-11-19 15:13:57.013594] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:06.934 [2024-11-19 15:13:57.013636] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:06.934 [2024-11-19 15:13:57.013645] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Raid, state offline 00:08:06.934 15:13:57 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@978 -- # wait 71601 00:08:07.194 [2024-11-19 15:13:57.319861] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:07.453 15:13:57 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@911 -- # return 0 00:08:07.453 00:08:07.453 real 0m2.442s 00:08:07.453 user 0m2.583s 00:08:07.453 sys 0m0.671s 00:08:07.453 15:13:57 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:07.453 ************************************ 00:08:07.453 END TEST raid0_resize_superblock_test 00:08:07.453 ************************************ 00:08:07.453 15:13:57 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.453 15:13:57 bdev_raid -- bdev/bdev_raid.sh@954 -- # run_test raid1_resize_superblock_test raid_resize_superblock_test 1 00:08:07.453 15:13:57 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:08:07.453 15:13:57 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:07.453 15:13:57 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:07.453 ************************************ 00:08:07.453 START TEST raid1_resize_superblock_test 00:08:07.453 ************************************ 00:08:07.453 15:13:57 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@1129 -- # raid_resize_superblock_test 1 00:08:07.453 15:13:57 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@854 -- # local raid_level=1 00:08:07.453 15:13:57 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@857 -- # raid_pid=71672 00:08:07.453 15:13:57 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@856 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:07.453 15:13:57 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@858 -- # echo 'Process raid pid: 71672' 00:08:07.453 Process raid pid: 71672 00:08:07.453 15:13:57 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@859 -- # waitforlisten 71672 00:08:07.453 15:13:57 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 71672 ']' 00:08:07.453 15:13:57 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:07.453 15:13:57 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:07.453 15:13:57 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:07.453 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:07.453 15:13:57 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:07.453 15:13:57 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.713 [2024-11-19 15:13:57.803131] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:08:07.713 [2024-11-19 15:13:57.803392] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:07.713 [2024-11-19 15:13:57.941367] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:07.713 [2024-11-19 15:13:57.986402] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:07.976 [2024-11-19 15:13:58.063013] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:07.976 [2024-11-19 15:13:58.063063] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:08.550 15:13:58 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:08.550 15:13:58 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:08:08.550 15:13:58 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@861 -- # rpc_cmd bdev_malloc_create -b malloc0 512 512 00:08:08.550 15:13:58 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:08.550 15:13:58 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.550 malloc0 00:08:08.550 15:13:58 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:08.550 15:13:58 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@863 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:08:08.550 15:13:58 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:08.550 15:13:58 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.550 [2024-11-19 15:13:58.865309] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:08:08.550 [2024-11-19 15:13:58.865473] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:08.550 [2024-11-19 15:13:58.865501] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:08:08.550 [2024-11-19 15:13:58.865513] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:08.550 [2024-11-19 15:13:58.868032] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:08.550 [2024-11-19 15:13:58.868074] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:08:08.550 pt0 00:08:08.550 15:13:58 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:08.550 15:13:58 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@864 -- # rpc_cmd bdev_lvol_create_lvstore pt0 lvs0 00:08:08.550 15:13:58 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:08.550 15:13:58 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.810 0899ac15-7d1f-4714-b86d-6312c74825ea 00:08:08.810 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:08.810 15:13:59 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@866 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol0 64 00:08:08.810 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:08.810 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.810 191105ed-f9b7-499d-9c8c-44f78d71a025 00:08:08.810 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:08.810 15:13:59 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@867 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol1 64 00:08:08.810 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:08.810 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.810 d15e49a4-0516-48b4-8ca2-fa27a3b07e96 00:08:08.810 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:08.810 15:13:59 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@869 -- # case $raid_level in 00:08:08.810 15:13:59 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@871 -- # rpc_cmd bdev_raid_create -n Raid -r 1 -b ''\''lvs0/lvol0 lvs0/lvol1'\''' -s 00:08:08.810 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:08.810 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.810 [2024-11-19 15:13:59.075952] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev 191105ed-f9b7-499d-9c8c-44f78d71a025 is claimed 00:08:08.810 [2024-11-19 15:13:59.076076] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev d15e49a4-0516-48b4-8ca2-fa27a3b07e96 is claimed 00:08:08.810 [2024-11-19 15:13:59.076193] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:08:08.810 [2024-11-19 15:13:59.076206] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 122880, blocklen 512 00:08:08.810 [2024-11-19 15:13:59.076509] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:08:08.810 [2024-11-19 15:13:59.076687] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:08:08.810 [2024-11-19 15:13:59.076698] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000001200 00:08:08.810 [2024-11-19 15:13:59.076837] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:08.810 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:08.810 15:13:59 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:08:08.810 15:13:59 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # jq '.[].num_blocks' 00:08:08.810 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:08.810 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.810 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:08.810 15:13:59 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # (( 64 == 64 )) 00:08:08.810 15:13:59 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:08:08.810 15:13:59 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # jq '.[].num_blocks' 00:08:08.810 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:08.810 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:09.071 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:09.071 15:13:59 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # (( 64 == 64 )) 00:08:09.071 15:13:59 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:08:09.071 15:13:59 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@881 -- # rpc_cmd bdev_get_bdevs -b Raid 00:08:09.071 15:13:59 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:08:09.071 15:13:59 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@881 -- # jq '.[].num_blocks' 00:08:09.071 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:09.071 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:09.071 [2024-11-19 15:13:59.192023] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:09.071 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:09.071 15:13:59 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:08:09.071 15:13:59 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:08:09.071 15:13:59 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@881 -- # (( 122880 == 122880 )) 00:08:09.071 15:13:59 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@885 -- # rpc_cmd bdev_lvol_resize lvs0/lvol0 100 00:08:09.071 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:09.071 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:09.071 [2024-11-19 15:13:59.235815] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:08:09.071 [2024-11-19 15:13:59.235843] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev '191105ed-f9b7-499d-9c8c-44f78d71a025' was resized: old size 131072, new size 204800 00:08:09.071 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:09.071 15:13:59 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@886 -- # rpc_cmd bdev_lvol_resize lvs0/lvol1 100 00:08:09.071 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:09.071 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:09.071 [2024-11-19 15:13:59.247782] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:08:09.071 [2024-11-19 15:13:59.247806] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'd15e49a4-0516-48b4-8ca2-fa27a3b07e96' was resized: old size 131072, new size 204800 00:08:09.071 [2024-11-19 15:13:59.247832] bdev_raid.c:2344:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 122880 to 196608 00:08:09.071 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:09.071 15:13:59 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:08:09.071 15:13:59 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # jq '.[].num_blocks' 00:08:09.071 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:09.071 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:09.071 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:09.071 15:13:59 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # (( 100 == 100 )) 00:08:09.071 15:13:59 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:08:09.071 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:09.071 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:09.071 15:13:59 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # jq '.[].num_blocks' 00:08:09.071 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:09.071 15:13:59 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # (( 100 == 100 )) 00:08:09.071 15:13:59 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:08:09.071 15:13:59 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@895 -- # jq '.[].num_blocks' 00:08:09.071 15:13:59 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:08:09.071 15:13:59 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@895 -- # rpc_cmd bdev_get_bdevs -b Raid 00:08:09.071 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:09.071 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:09.071 [2024-11-19 15:13:59.359646] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:09.071 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:09.071 15:13:59 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:08:09.071 15:13:59 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:08:09.071 15:13:59 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@895 -- # (( 196608 == 196608 )) 00:08:09.071 15:13:59 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@898 -- # rpc_cmd bdev_passthru_delete pt0 00:08:09.071 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:09.072 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:09.072 [2024-11-19 15:13:59.403377] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev pt0 being removed: closing lvstore lvs0 00:08:09.072 [2024-11-19 15:13:59.403444] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol0 00:08:09.072 [2024-11-19 15:13:59.403469] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol1 00:08:09.072 [2024-11-19 15:13:59.403613] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:09.072 [2024-11-19 15:13:59.403757] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:09.072 [2024-11-19 15:13:59.403809] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:09.072 [2024-11-19 15:13:59.403822] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Raid, state offline 00:08:09.072 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:09.072 15:13:59 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@899 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:08:09.072 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:09.072 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:09.346 [2024-11-19 15:13:59.411332] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:08:09.346 [2024-11-19 15:13:59.411380] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:09.346 [2024-11-19 15:13:59.411396] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:08:09.346 [2024-11-19 15:13:59.411407] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:09.346 [2024-11-19 15:13:59.413907] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:09.346 [2024-11-19 15:13:59.414037] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:08:09.346 pt0 00:08:09.346 [2024-11-19 15:13:59.415384] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev 191105ed-f9b7-499d-9c8c-44f78d71a025 00:08:09.346 [2024-11-19 15:13:59.415451] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev 191105ed-f9b7-499d-9c8c-44f78d71a025 is claimed 00:08:09.346 [2024-11-19 15:13:59.415525] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev d15e49a4-0516-48b4-8ca2-fa27a3b07e96 00:08:09.346 [2024-11-19 15:13:59.415546] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev d15e49a4-0516-48b4-8ca2-fa27a3b07e96 is claimed 00:08:09.346 [2024-11-19 15:13:59.415676] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev d15e49a4-0516-48b4-8ca2-fa27a3b07e96 (2) smaller than existing raid bdev Raid (3) 00:08:09.346 [2024-11-19 15:13:59.415705] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev 191105ed-f9b7-499d-9c8c-44f78d71a025: File exists 00:08:09.346 [2024-11-19 15:13:59.415768] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:08:09.346 [2024-11-19 15:13:59.415778] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:08:09.346 [2024-11-19 15:13:59.416049] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:08:09.346 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:09.346 [2024-11-19 15:13:59.416232] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:08:09.346 [2024-11-19 15:13:59.416244] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000001580 00:08:09.346 15:13:59 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@900 -- # rpc_cmd bdev_wait_for_examine 00:08:09.346 [2024-11-19 15:13:59.416360] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:09.346 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:09.346 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:09.346 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:09.346 15:13:59 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:08:09.346 15:13:59 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@906 -- # rpc_cmd bdev_get_bdevs -b Raid 00:08:09.346 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:09.346 15:13:59 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:08:09.346 15:13:59 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@906 -- # jq '.[].num_blocks' 00:08:09.346 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:09.346 [2024-11-19 15:13:59.435630] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:09.346 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:09.346 15:13:59 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:08:09.346 15:13:59 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:08:09.346 15:13:59 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@906 -- # (( 196608 == 196608 )) 00:08:09.346 15:13:59 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@909 -- # killprocess 71672 00:08:09.346 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 71672 ']' 00:08:09.346 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@958 -- # kill -0 71672 00:08:09.346 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@959 -- # uname 00:08:09.346 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:09.346 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71672 00:08:09.346 killing process with pid 71672 00:08:09.346 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:09.346 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:09.346 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71672' 00:08:09.346 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@973 -- # kill 71672 00:08:09.346 [2024-11-19 15:13:59.510427] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:09.346 [2024-11-19 15:13:59.510480] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:09.346 [2024-11-19 15:13:59.510530] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:09.346 [2024-11-19 15:13:59.510540] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Raid, state offline 00:08:09.346 15:13:59 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@978 -- # wait 71672 00:08:09.607 [2024-11-19 15:13:59.815621] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:09.866 ************************************ 00:08:09.866 END TEST raid1_resize_superblock_test 00:08:09.866 ************************************ 00:08:09.866 15:14:00 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@911 -- # return 0 00:08:09.866 00:08:09.866 real 0m2.426s 00:08:09.866 user 0m2.613s 00:08:09.866 sys 0m0.625s 00:08:09.866 15:14:00 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:09.866 15:14:00 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:09.866 15:14:00 bdev_raid -- bdev/bdev_raid.sh@956 -- # uname -s 00:08:09.866 15:14:00 bdev_raid -- bdev/bdev_raid.sh@956 -- # '[' Linux = Linux ']' 00:08:09.866 15:14:00 bdev_raid -- bdev/bdev_raid.sh@956 -- # modprobe -n nbd 00:08:10.126 15:14:00 bdev_raid -- bdev/bdev_raid.sh@957 -- # has_nbd=true 00:08:10.126 15:14:00 bdev_raid -- bdev/bdev_raid.sh@958 -- # modprobe nbd 00:08:10.126 15:14:00 bdev_raid -- bdev/bdev_raid.sh@959 -- # run_test raid_function_test_raid0 raid_function_test raid0 00:08:10.126 15:14:00 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:08:10.126 15:14:00 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:10.126 15:14:00 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:10.126 ************************************ 00:08:10.126 START TEST raid_function_test_raid0 00:08:10.126 ************************************ 00:08:10.126 15:14:00 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@1129 -- # raid_function_test raid0 00:08:10.126 15:14:00 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@64 -- # local raid_level=raid0 00:08:10.126 15:14:00 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@65 -- # local nbd=/dev/nbd0 00:08:10.126 15:14:00 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@66 -- # local raid_bdev 00:08:10.126 15:14:00 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@69 -- # raid_pid=71753 00:08:10.126 Process raid pid: 71753 00:08:10.126 15:14:00 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@68 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:10.126 15:14:00 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@70 -- # echo 'Process raid pid: 71753' 00:08:10.126 15:14:00 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@71 -- # waitforlisten 71753 00:08:10.126 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:10.126 15:14:00 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@835 -- # '[' -z 71753 ']' 00:08:10.126 15:14:00 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:10.126 15:14:00 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:10.126 15:14:00 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:10.126 15:14:00 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:10.126 15:14:00 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:08:10.126 [2024-11-19 15:14:00.300845] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:08:10.126 [2024-11-19 15:14:00.301085] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:10.127 [2024-11-19 15:14:00.458899] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:10.386 [2024-11-19 15:14:00.497943] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:10.386 [2024-11-19 15:14:00.574333] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:10.387 [2024-11-19 15:14:00.574378] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:10.956 15:14:01 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:10.956 15:14:01 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@868 -- # return 0 00:08:10.956 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@73 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_1 00:08:10.956 15:14:01 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:10.956 15:14:01 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:08:10.956 Base_1 00:08:10.956 15:14:01 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:10.956 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@74 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_2 00:08:10.956 15:14:01 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:10.956 15:14:01 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:08:10.956 Base_2 00:08:10.956 15:14:01 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:10.956 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@75 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''Base_1 Base_2'\''' -n raid 00:08:10.956 15:14:01 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:10.956 15:14:01 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:08:10.956 [2024-11-19 15:14:01.183747] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:08:10.956 [2024-11-19 15:14:01.186000] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:08:10.956 [2024-11-19 15:14:01.186065] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:08:10.956 [2024-11-19 15:14:01.186078] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:08:10.956 [2024-11-19 15:14:01.186362] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:08:10.957 [2024-11-19 15:14:01.186500] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:08:10.957 [2024-11-19 15:14:01.186514] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid, raid_bdev 0x617000001200 00:08:10.957 [2024-11-19 15:14:01.186640] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:10.957 15:14:01 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:10.957 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@77 -- # rpc_cmd bdev_raid_get_bdevs online 00:08:10.957 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@77 -- # jq -r '.[0]["name"] | select(.)' 00:08:10.957 15:14:01 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:10.957 15:14:01 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:08:10.957 15:14:01 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:10.957 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@77 -- # raid_bdev=raid 00:08:10.957 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@78 -- # '[' raid = '' ']' 00:08:10.957 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@83 -- # nbd_start_disks /var/tmp/spdk.sock raid /dev/nbd0 00:08:10.957 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:08:10.957 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@10 -- # bdev_list=('raid') 00:08:10.957 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@10 -- # local bdev_list 00:08:10.957 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:08:10.957 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@11 -- # local nbd_list 00:08:10.957 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@12 -- # local i 00:08:10.957 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:08:10.957 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:08:10.957 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid /dev/nbd0 00:08:11.216 [2024-11-19 15:14:01.427290] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:08:11.216 /dev/nbd0 00:08:11.216 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:08:11.216 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:08:11.216 15:14:01 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:08:11.216 15:14:01 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@873 -- # local i 00:08:11.216 15:14:01 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:08:11.216 15:14:01 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:08:11.216 15:14:01 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:08:11.216 15:14:01 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@877 -- # break 00:08:11.216 15:14:01 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:08:11.216 15:14:01 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:08:11.216 15:14:01 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:08:11.216 1+0 records in 00:08:11.216 1+0 records out 00:08:11.216 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000559494 s, 7.3 MB/s 00:08:11.216 15:14:01 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:08:11.216 15:14:01 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@890 -- # size=4096 00:08:11.216 15:14:01 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:08:11.216 15:14:01 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:08:11.216 15:14:01 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@893 -- # return 0 00:08:11.216 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:08:11.216 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:08:11.216 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@84 -- # nbd_get_count /var/tmp/spdk.sock 00:08:11.216 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:08:11.216 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:08:11.476 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:08:11.476 { 00:08:11.476 "nbd_device": "/dev/nbd0", 00:08:11.476 "bdev_name": "raid" 00:08:11.476 } 00:08:11.476 ]' 00:08:11.476 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:08:11.476 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # echo '[ 00:08:11.476 { 00:08:11.476 "nbd_device": "/dev/nbd0", 00:08:11.476 "bdev_name": "raid" 00:08:11.476 } 00:08:11.476 ]' 00:08:11.476 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # nbd_disks_name=/dev/nbd0 00:08:11.476 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:08:11.476 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # echo /dev/nbd0 00:08:11.476 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # count=1 00:08:11.476 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@66 -- # echo 1 00:08:11.476 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@84 -- # count=1 00:08:11.476 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@85 -- # '[' 1 -ne 1 ']' 00:08:11.476 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@89 -- # raid_unmap_data_verify /dev/nbd0 00:08:11.476 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@17 -- # hash blkdiscard 00:08:11.476 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@18 -- # local nbd=/dev/nbd0 00:08:11.476 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@19 -- # local blksize 00:08:11.476 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # lsblk -o LOG-SEC /dev/nbd0 00:08:11.476 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # cut -d ' ' -f 5 00:08:11.476 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # grep -v LOG-SEC 00:08:11.476 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # blksize=512 00:08:11.476 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@21 -- # local rw_blk_num=4096 00:08:11.476 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@22 -- # local rw_len=2097152 00:08:11.476 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@23 -- # unmap_blk_offs=('0' '1028' '321') 00:08:11.476 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@23 -- # local unmap_blk_offs 00:08:11.476 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@24 -- # unmap_blk_nums=('128' '2035' '456') 00:08:11.476 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@24 -- # local unmap_blk_nums 00:08:11.476 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@25 -- # local unmap_off 00:08:11.476 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@26 -- # local unmap_len 00:08:11.476 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@29 -- # dd if=/dev/urandom of=/raidtest/raidrandtest bs=512 count=4096 00:08:11.476 4096+0 records in 00:08:11.476 4096+0 records out 00:08:11.476 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.0357731 s, 58.6 MB/s 00:08:11.476 15:14:01 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@30 -- # dd if=/raidtest/raidrandtest of=/dev/nbd0 bs=512 count=4096 oflag=direct 00:08:11.735 4096+0 records in 00:08:11.735 4096+0 records out 00:08:11.735 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.209117 s, 10.0 MB/s 00:08:11.735 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@31 -- # blockdev --flushbufs /dev/nbd0 00:08:11.735 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@34 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:08:11.735 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i = 0 )) 00:08:11.735 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:08:11.735 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@37 -- # unmap_off=0 00:08:11.735 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@38 -- # unmap_len=65536 00:08:11.735 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=0 count=128 conv=notrunc 00:08:11.735 128+0 records in 00:08:11.735 128+0 records out 00:08:11.735 65536 bytes (66 kB, 64 KiB) copied, 0.00116528 s, 56.2 MB/s 00:08:11.735 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 0 -l 65536 /dev/nbd0 00:08:11.735 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:08:11.735 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:08:11.735 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:08:11.735 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:08:11.736 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@37 -- # unmap_off=526336 00:08:11.736 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@38 -- # unmap_len=1041920 00:08:11.736 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=1028 count=2035 conv=notrunc 00:08:11.996 2035+0 records in 00:08:11.996 2035+0 records out 00:08:11.996 1041920 bytes (1.0 MB, 1018 KiB) copied, 0.0141142 s, 73.8 MB/s 00:08:11.996 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 526336 -l 1041920 /dev/nbd0 00:08:11.996 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:08:11.996 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:08:11.996 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:08:11.996 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:08:11.996 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@37 -- # unmap_off=164352 00:08:11.996 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@38 -- # unmap_len=233472 00:08:11.996 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=321 count=456 conv=notrunc 00:08:11.996 456+0 records in 00:08:11.996 456+0 records out 00:08:11.996 233472 bytes (233 kB, 228 KiB) copied, 0.00393411 s, 59.3 MB/s 00:08:11.996 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 164352 -l 233472 /dev/nbd0 00:08:11.996 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:08:11.996 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:08:11.996 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:08:11.996 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:08:11.996 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@52 -- # return 0 00:08:11.996 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@91 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:08:11.996 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:08:11.996 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:08:11.996 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@50 -- # local nbd_list 00:08:11.996 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@51 -- # local i 00:08:11.996 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:08:11.996 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:08:12.256 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:08:12.256 [2024-11-19 15:14:02.340019] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:12.256 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:08:12.256 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:08:12.256 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:08:12.256 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:08:12.256 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:08:12.256 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@41 -- # break 00:08:12.256 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@45 -- # return 0 00:08:12.256 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@92 -- # nbd_get_count /var/tmp/spdk.sock 00:08:12.256 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:08:12.256 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:08:12.256 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:08:12.256 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # echo '[]' 00:08:12.256 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:08:12.516 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:08:12.516 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:08:12.516 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # echo '' 00:08:12.516 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # true 00:08:12.516 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # count=0 00:08:12.517 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@66 -- # echo 0 00:08:12.517 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@92 -- # count=0 00:08:12.517 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@93 -- # '[' 0 -ne 0 ']' 00:08:12.517 15:14:02 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@97 -- # killprocess 71753 00:08:12.517 15:14:02 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@954 -- # '[' -z 71753 ']' 00:08:12.517 15:14:02 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@958 -- # kill -0 71753 00:08:12.517 15:14:02 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@959 -- # uname 00:08:12.517 15:14:02 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:12.517 15:14:02 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71753 00:08:12.517 killing process with pid 71753 00:08:12.517 15:14:02 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:12.517 15:14:02 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:12.517 15:14:02 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71753' 00:08:12.517 15:14:02 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@973 -- # kill 71753 00:08:12.517 [2024-11-19 15:14:02.665538] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:12.517 [2024-11-19 15:14:02.665641] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:12.517 [2024-11-19 15:14:02.665694] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:12.517 [2024-11-19 15:14:02.665710] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid, state offline 00:08:12.517 15:14:02 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@978 -- # wait 71753 00:08:12.517 [2024-11-19 15:14:02.705605] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:12.776 15:14:03 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@99 -- # return 0 00:08:12.776 00:08:12.776 real 0m2.801s 00:08:12.776 user 0m3.345s 00:08:12.776 sys 0m1.022s 00:08:12.776 ************************************ 00:08:12.776 END TEST raid_function_test_raid0 00:08:12.776 ************************************ 00:08:12.776 15:14:03 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:12.776 15:14:03 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:08:12.776 15:14:03 bdev_raid -- bdev/bdev_raid.sh@960 -- # run_test raid_function_test_concat raid_function_test concat 00:08:12.776 15:14:03 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:08:12.776 15:14:03 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:12.776 15:14:03 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:12.776 ************************************ 00:08:12.776 START TEST raid_function_test_concat 00:08:12.776 ************************************ 00:08:12.776 15:14:03 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@1129 -- # raid_function_test concat 00:08:12.776 15:14:03 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@64 -- # local raid_level=concat 00:08:12.776 15:14:03 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@65 -- # local nbd=/dev/nbd0 00:08:12.776 15:14:03 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@66 -- # local raid_bdev 00:08:12.776 Process raid pid: 71873 00:08:12.776 15:14:03 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@69 -- # raid_pid=71873 00:08:12.776 15:14:03 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@68 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:12.776 15:14:03 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@70 -- # echo 'Process raid pid: 71873' 00:08:12.776 15:14:03 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@71 -- # waitforlisten 71873 00:08:12.776 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:12.776 15:14:03 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@835 -- # '[' -z 71873 ']' 00:08:12.776 15:14:03 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:12.777 15:14:03 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:12.777 15:14:03 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:12.777 15:14:03 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:12.777 15:14:03 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:08:13.037 [2024-11-19 15:14:03.179311] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:08:13.037 [2024-11-19 15:14:03.179533] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:13.037 [2024-11-19 15:14:03.335132] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:13.037 [2024-11-19 15:14:03.373266] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:13.296 [2024-11-19 15:14:03.449522] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:13.296 [2024-11-19 15:14:03.449667] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:13.866 15:14:04 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:13.867 15:14:04 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@868 -- # return 0 00:08:13.867 15:14:04 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@73 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_1 00:08:13.867 15:14:04 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:13.867 15:14:04 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:08:13.867 Base_1 00:08:13.867 15:14:04 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:13.867 15:14:04 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@74 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_2 00:08:13.867 15:14:04 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:13.867 15:14:04 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:08:13.867 Base_2 00:08:13.867 15:14:04 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:13.867 15:14:04 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@75 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''Base_1 Base_2'\''' -n raid 00:08:13.867 15:14:04 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:13.867 15:14:04 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:08:13.867 [2024-11-19 15:14:04.083199] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:08:13.867 [2024-11-19 15:14:04.085464] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:08:13.867 [2024-11-19 15:14:04.085655] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:08:13.867 [2024-11-19 15:14:04.085675] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:08:13.867 [2024-11-19 15:14:04.086025] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:08:13.867 [2024-11-19 15:14:04.086191] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:08:13.867 [2024-11-19 15:14:04.086220] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid, raid_bdev 0x617000001200 00:08:13.867 [2024-11-19 15:14:04.086389] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:13.867 15:14:04 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:13.867 15:14:04 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@77 -- # rpc_cmd bdev_raid_get_bdevs online 00:08:13.867 15:14:04 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@77 -- # jq -r '.[0]["name"] | select(.)' 00:08:13.867 15:14:04 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:13.867 15:14:04 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:08:13.867 15:14:04 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:13.867 15:14:04 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@77 -- # raid_bdev=raid 00:08:13.867 15:14:04 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@78 -- # '[' raid = '' ']' 00:08:13.867 15:14:04 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@83 -- # nbd_start_disks /var/tmp/spdk.sock raid /dev/nbd0 00:08:13.867 15:14:04 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:08:13.867 15:14:04 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@10 -- # bdev_list=('raid') 00:08:13.867 15:14:04 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:08:13.867 15:14:04 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:08:13.867 15:14:04 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:08:13.867 15:14:04 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@12 -- # local i 00:08:13.867 15:14:04 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:08:13.867 15:14:04 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:08:13.867 15:14:04 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid /dev/nbd0 00:08:14.127 [2024-11-19 15:14:04.326896] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:08:14.127 /dev/nbd0 00:08:14.127 15:14:04 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:08:14.127 15:14:04 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:08:14.127 15:14:04 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:08:14.127 15:14:04 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@873 -- # local i 00:08:14.127 15:14:04 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:08:14.127 15:14:04 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:08:14.127 15:14:04 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:08:14.127 15:14:04 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@877 -- # break 00:08:14.127 15:14:04 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:08:14.127 15:14:04 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:08:14.127 15:14:04 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:08:14.127 1+0 records in 00:08:14.127 1+0 records out 00:08:14.127 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000358981 s, 11.4 MB/s 00:08:14.127 15:14:04 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:08:14.127 15:14:04 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@890 -- # size=4096 00:08:14.127 15:14:04 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:08:14.127 15:14:04 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:08:14.127 15:14:04 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@893 -- # return 0 00:08:14.127 15:14:04 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:08:14.127 15:14:04 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:08:14.127 15:14:04 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@84 -- # nbd_get_count /var/tmp/spdk.sock 00:08:14.127 15:14:04 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:08:14.127 15:14:04 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:08:14.387 15:14:04 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:08:14.387 { 00:08:14.387 "nbd_device": "/dev/nbd0", 00:08:14.387 "bdev_name": "raid" 00:08:14.387 } 00:08:14.387 ]' 00:08:14.387 15:14:04 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # echo '[ 00:08:14.387 { 00:08:14.387 "nbd_device": "/dev/nbd0", 00:08:14.387 "bdev_name": "raid" 00:08:14.387 } 00:08:14.387 ]' 00:08:14.387 15:14:04 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:08:14.387 15:14:04 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # nbd_disks_name=/dev/nbd0 00:08:14.387 15:14:04 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # echo /dev/nbd0 00:08:14.387 15:14:04 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:08:14.387 15:14:04 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # count=1 00:08:14.387 15:14:04 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@66 -- # echo 1 00:08:14.387 15:14:04 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@84 -- # count=1 00:08:14.387 15:14:04 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@85 -- # '[' 1 -ne 1 ']' 00:08:14.387 15:14:04 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@89 -- # raid_unmap_data_verify /dev/nbd0 00:08:14.387 15:14:04 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@17 -- # hash blkdiscard 00:08:14.387 15:14:04 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@18 -- # local nbd=/dev/nbd0 00:08:14.387 15:14:04 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@19 -- # local blksize 00:08:14.387 15:14:04 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # lsblk -o LOG-SEC /dev/nbd0 00:08:14.387 15:14:04 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # grep -v LOG-SEC 00:08:14.387 15:14:04 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # cut -d ' ' -f 5 00:08:14.387 15:14:04 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # blksize=512 00:08:14.387 15:14:04 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@21 -- # local rw_blk_num=4096 00:08:14.387 15:14:04 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@22 -- # local rw_len=2097152 00:08:14.387 15:14:04 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@23 -- # unmap_blk_offs=('0' '1028' '321') 00:08:14.387 15:14:04 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@23 -- # local unmap_blk_offs 00:08:14.387 15:14:04 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@24 -- # unmap_blk_nums=('128' '2035' '456') 00:08:14.387 15:14:04 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@24 -- # local unmap_blk_nums 00:08:14.387 15:14:04 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@25 -- # local unmap_off 00:08:14.387 15:14:04 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@26 -- # local unmap_len 00:08:14.388 15:14:04 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@29 -- # dd if=/dev/urandom of=/raidtest/raidrandtest bs=512 count=4096 00:08:14.388 4096+0 records in 00:08:14.388 4096+0 records out 00:08:14.388 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.0317926 s, 66.0 MB/s 00:08:14.388 15:14:04 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@30 -- # dd if=/raidtest/raidrandtest of=/dev/nbd0 bs=512 count=4096 oflag=direct 00:08:14.648 4096+0 records in 00:08:14.648 4096+0 records out 00:08:14.648 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.226998 s, 9.2 MB/s 00:08:14.648 15:14:04 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@31 -- # blockdev --flushbufs /dev/nbd0 00:08:14.648 15:14:04 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@34 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:08:14.648 15:14:04 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i = 0 )) 00:08:14.648 15:14:04 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:08:14.648 15:14:04 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@37 -- # unmap_off=0 00:08:14.648 15:14:04 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@38 -- # unmap_len=65536 00:08:14.648 15:14:04 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=0 count=128 conv=notrunc 00:08:14.648 128+0 records in 00:08:14.648 128+0 records out 00:08:14.648 65536 bytes (66 kB, 64 KiB) copied, 0.00107382 s, 61.0 MB/s 00:08:14.648 15:14:04 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 0 -l 65536 /dev/nbd0 00:08:14.648 15:14:04 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:08:14.648 15:14:04 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:08:14.648 15:14:04 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:08:14.648 15:14:04 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:08:14.648 15:14:04 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@37 -- # unmap_off=526336 00:08:14.648 15:14:04 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@38 -- # unmap_len=1041920 00:08:14.648 15:14:04 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=1028 count=2035 conv=notrunc 00:08:14.908 2035+0 records in 00:08:14.908 2035+0 records out 00:08:14.908 1041920 bytes (1.0 MB, 1018 KiB) copied, 0.0147124 s, 70.8 MB/s 00:08:14.908 15:14:05 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 526336 -l 1041920 /dev/nbd0 00:08:14.908 15:14:05 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:08:14.908 15:14:05 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:08:14.908 15:14:05 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:08:14.908 15:14:05 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:08:14.908 15:14:05 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@37 -- # unmap_off=164352 00:08:14.908 15:14:05 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@38 -- # unmap_len=233472 00:08:14.908 15:14:05 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=321 count=456 conv=notrunc 00:08:14.908 456+0 records in 00:08:14.908 456+0 records out 00:08:14.908 233472 bytes (233 kB, 228 KiB) copied, 0.00361224 s, 64.6 MB/s 00:08:14.908 15:14:05 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 164352 -l 233472 /dev/nbd0 00:08:14.908 15:14:05 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:08:14.908 15:14:05 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:08:14.908 15:14:05 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:08:14.908 15:14:05 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:08:14.908 15:14:05 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@52 -- # return 0 00:08:14.908 15:14:05 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@91 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:08:14.908 15:14:05 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:08:14.908 15:14:05 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:08:14.908 15:14:05 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:08:14.909 15:14:05 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@51 -- # local i 00:08:14.909 15:14:05 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:08:14.909 15:14:05 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:08:15.169 15:14:05 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:08:15.169 [2024-11-19 15:14:05.270469] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:15.169 15:14:05 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:08:15.169 15:14:05 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:08:15.169 15:14:05 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:08:15.169 15:14:05 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:08:15.169 15:14:05 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:08:15.169 15:14:05 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@41 -- # break 00:08:15.169 15:14:05 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@45 -- # return 0 00:08:15.169 15:14:05 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@92 -- # nbd_get_count /var/tmp/spdk.sock 00:08:15.169 15:14:05 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:08:15.169 15:14:05 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:08:15.169 15:14:05 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:08:15.169 15:14:05 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:08:15.169 15:14:05 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:08:15.429 15:14:05 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:08:15.429 15:14:05 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # echo '' 00:08:15.429 15:14:05 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:08:15.429 15:14:05 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # true 00:08:15.429 15:14:05 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # count=0 00:08:15.429 15:14:05 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@66 -- # echo 0 00:08:15.429 15:14:05 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@92 -- # count=0 00:08:15.429 15:14:05 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@93 -- # '[' 0 -ne 0 ']' 00:08:15.429 15:14:05 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@97 -- # killprocess 71873 00:08:15.429 15:14:05 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@954 -- # '[' -z 71873 ']' 00:08:15.429 15:14:05 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@958 -- # kill -0 71873 00:08:15.429 15:14:05 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@959 -- # uname 00:08:15.429 15:14:05 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:15.429 15:14:05 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71873 00:08:15.429 killing process with pid 71873 00:08:15.429 15:14:05 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:15.429 15:14:05 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:15.429 15:14:05 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71873' 00:08:15.429 15:14:05 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@973 -- # kill 71873 00:08:15.429 [2024-11-19 15:14:05.569763] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:15.429 [2024-11-19 15:14:05.569896] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:15.429 15:14:05 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@978 -- # wait 71873 00:08:15.430 [2024-11-19 15:14:05.569961] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:15.430 [2024-11-19 15:14:05.569977] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid, state offline 00:08:15.430 [2024-11-19 15:14:05.610928] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:15.689 15:14:05 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@99 -- # return 0 00:08:15.689 ************************************ 00:08:15.689 END TEST raid_function_test_concat 00:08:15.689 ************************************ 00:08:15.689 00:08:15.689 real 0m2.836s 00:08:15.689 user 0m3.395s 00:08:15.690 sys 0m1.009s 00:08:15.690 15:14:05 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:15.690 15:14:05 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:08:15.690 15:14:05 bdev_raid -- bdev/bdev_raid.sh@963 -- # run_test raid0_resize_test raid_resize_test 0 00:08:15.690 15:14:05 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:08:15.690 15:14:05 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:15.690 15:14:05 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:15.690 ************************************ 00:08:15.690 START TEST raid0_resize_test 00:08:15.690 ************************************ 00:08:15.690 15:14:06 bdev_raid.raid0_resize_test -- common/autotest_common.sh@1129 -- # raid_resize_test 0 00:08:15.690 15:14:06 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@332 -- # local raid_level=0 00:08:15.690 15:14:06 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@333 -- # local blksize=512 00:08:15.690 15:14:06 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@334 -- # local bdev_size_mb=32 00:08:15.690 15:14:06 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@335 -- # local new_bdev_size_mb=64 00:08:15.690 15:14:06 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@336 -- # local blkcnt 00:08:15.690 15:14:06 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@337 -- # local raid_size_mb 00:08:15.690 15:14:06 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@338 -- # local new_raid_size_mb 00:08:15.690 15:14:06 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@339 -- # local expected_size 00:08:15.690 15:14:06 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@342 -- # raid_pid=71984 00:08:15.690 15:14:06 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@341 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:15.690 15:14:06 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@343 -- # echo 'Process raid pid: 71984' 00:08:15.690 Process raid pid: 71984 00:08:15.690 15:14:06 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@344 -- # waitforlisten 71984 00:08:15.690 15:14:06 bdev_raid.raid0_resize_test -- common/autotest_common.sh@835 -- # '[' -z 71984 ']' 00:08:15.690 15:14:06 bdev_raid.raid0_resize_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:15.690 15:14:06 bdev_raid.raid0_resize_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:15.690 15:14:06 bdev_raid.raid0_resize_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:15.690 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:15.690 15:14:06 bdev_raid.raid0_resize_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:15.690 15:14:06 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.950 [2024-11-19 15:14:06.090514] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:08:15.950 [2024-11-19 15:14:06.090723] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:15.950 [2024-11-19 15:14:06.247316] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:15.950 [2024-11-19 15:14:06.285365] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:16.210 [2024-11-19 15:14:06.361417] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:16.210 [2024-11-19 15:14:06.361463] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:16.779 15:14:06 bdev_raid.raid0_resize_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:16.779 15:14:06 bdev_raid.raid0_resize_test -- common/autotest_common.sh@868 -- # return 0 00:08:16.779 15:14:06 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@346 -- # rpc_cmd bdev_null_create Base_1 32 512 00:08:16.780 15:14:06 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:16.780 15:14:06 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.780 Base_1 00:08:16.780 15:14:06 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:16.780 15:14:06 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@347 -- # rpc_cmd bdev_null_create Base_2 32 512 00:08:16.780 15:14:06 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:16.780 15:14:06 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.780 Base_2 00:08:16.780 15:14:06 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:16.780 15:14:06 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@349 -- # '[' 0 -eq 0 ']' 00:08:16.780 15:14:06 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@350 -- # rpc_cmd bdev_raid_create -z 64 -r 0 -b ''\''Base_1 Base_2'\''' -n Raid 00:08:16.780 15:14:06 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:16.780 15:14:06 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.780 [2024-11-19 15:14:06.948326] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:08:16.780 [2024-11-19 15:14:06.950504] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:08:16.780 [2024-11-19 15:14:06.950559] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:08:16.780 [2024-11-19 15:14:06.950570] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:08:16.780 [2024-11-19 15:14:06.950852] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000021f0 00:08:16.780 [2024-11-19 15:14:06.950971] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:08:16.780 [2024-11-19 15:14:06.951001] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000001200 00:08:16.780 [2024-11-19 15:14:06.951120] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:16.780 15:14:06 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:16.780 15:14:06 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@356 -- # rpc_cmd bdev_null_resize Base_1 64 00:08:16.780 15:14:06 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:16.780 15:14:06 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.780 [2024-11-19 15:14:06.960270] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:08:16.780 [2024-11-19 15:14:06.960368] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_1' was resized: old size 65536, new size 131072 00:08:16.780 true 00:08:16.780 15:14:06 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:16.780 15:14:06 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@359 -- # rpc_cmd bdev_get_bdevs -b Raid 00:08:16.780 15:14:06 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:16.780 15:14:06 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.780 15:14:06 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@359 -- # jq '.[].num_blocks' 00:08:16.780 [2024-11-19 15:14:06.972453] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:16.780 15:14:06 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:16.780 15:14:07 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@359 -- # blkcnt=131072 00:08:16.780 15:14:07 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@360 -- # raid_size_mb=64 00:08:16.780 15:14:07 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@361 -- # '[' 0 -eq 0 ']' 00:08:16.780 15:14:07 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@362 -- # expected_size=64 00:08:16.780 15:14:07 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@366 -- # '[' 64 '!=' 64 ']' 00:08:16.780 15:14:07 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@372 -- # rpc_cmd bdev_null_resize Base_2 64 00:08:16.780 15:14:07 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:16.780 15:14:07 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.780 [2024-11-19 15:14:07.024129] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:08:16.780 [2024-11-19 15:14:07.024151] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_2' was resized: old size 65536, new size 131072 00:08:16.780 [2024-11-19 15:14:07.024178] bdev_raid.c:2344:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 131072 to 262144 00:08:16.780 true 00:08:16.780 15:14:07 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:16.780 15:14:07 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@375 -- # rpc_cmd bdev_get_bdevs -b Raid 00:08:16.780 15:14:07 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:16.780 15:14:07 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.780 15:14:07 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@375 -- # jq '.[].num_blocks' 00:08:16.780 [2024-11-19 15:14:07.036301] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:16.780 15:14:07 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:16.780 15:14:07 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@375 -- # blkcnt=262144 00:08:16.780 15:14:07 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@376 -- # raid_size_mb=128 00:08:16.780 15:14:07 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@377 -- # '[' 0 -eq 0 ']' 00:08:16.780 15:14:07 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@378 -- # expected_size=128 00:08:16.780 15:14:07 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@382 -- # '[' 128 '!=' 128 ']' 00:08:16.780 15:14:07 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@387 -- # killprocess 71984 00:08:16.780 15:14:07 bdev_raid.raid0_resize_test -- common/autotest_common.sh@954 -- # '[' -z 71984 ']' 00:08:16.780 15:14:07 bdev_raid.raid0_resize_test -- common/autotest_common.sh@958 -- # kill -0 71984 00:08:16.780 15:14:07 bdev_raid.raid0_resize_test -- common/autotest_common.sh@959 -- # uname 00:08:16.780 15:14:07 bdev_raid.raid0_resize_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:16.780 15:14:07 bdev_raid.raid0_resize_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71984 00:08:17.040 15:14:07 bdev_raid.raid0_resize_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:17.040 15:14:07 bdev_raid.raid0_resize_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:17.040 15:14:07 bdev_raid.raid0_resize_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71984' 00:08:17.040 killing process with pid 71984 00:08:17.040 15:14:07 bdev_raid.raid0_resize_test -- common/autotest_common.sh@973 -- # kill 71984 00:08:17.040 [2024-11-19 15:14:07.126792] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:17.040 [2024-11-19 15:14:07.126924] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:17.040 [2024-11-19 15:14:07.126999] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:17.040 15:14:07 bdev_raid.raid0_resize_test -- common/autotest_common.sh@978 -- # wait 71984 00:08:17.040 [2024-11-19 15:14:07.127095] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Raid, state offline 00:08:17.040 [2024-11-19 15:14:07.129237] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:17.300 15:14:07 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@389 -- # return 0 00:08:17.300 00:08:17.300 real 0m1.447s 00:08:17.300 user 0m1.559s 00:08:17.300 sys 0m0.359s 00:08:17.300 15:14:07 bdev_raid.raid0_resize_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:17.300 15:14:07 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.300 ************************************ 00:08:17.300 END TEST raid0_resize_test 00:08:17.300 ************************************ 00:08:17.300 15:14:07 bdev_raid -- bdev/bdev_raid.sh@964 -- # run_test raid1_resize_test raid_resize_test 1 00:08:17.300 15:14:07 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:08:17.300 15:14:07 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:17.300 15:14:07 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:17.300 ************************************ 00:08:17.300 START TEST raid1_resize_test 00:08:17.300 ************************************ 00:08:17.300 15:14:07 bdev_raid.raid1_resize_test -- common/autotest_common.sh@1129 -- # raid_resize_test 1 00:08:17.300 15:14:07 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@332 -- # local raid_level=1 00:08:17.300 15:14:07 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@333 -- # local blksize=512 00:08:17.300 15:14:07 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@334 -- # local bdev_size_mb=32 00:08:17.300 15:14:07 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@335 -- # local new_bdev_size_mb=64 00:08:17.300 15:14:07 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@336 -- # local blkcnt 00:08:17.300 15:14:07 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@337 -- # local raid_size_mb 00:08:17.300 15:14:07 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@338 -- # local new_raid_size_mb 00:08:17.300 15:14:07 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@339 -- # local expected_size 00:08:17.300 Process raid pid: 72035 00:08:17.300 15:14:07 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@342 -- # raid_pid=72035 00:08:17.300 15:14:07 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@341 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:17.300 15:14:07 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@343 -- # echo 'Process raid pid: 72035' 00:08:17.300 15:14:07 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@344 -- # waitforlisten 72035 00:08:17.300 15:14:07 bdev_raid.raid1_resize_test -- common/autotest_common.sh@835 -- # '[' -z 72035 ']' 00:08:17.300 15:14:07 bdev_raid.raid1_resize_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:17.300 15:14:07 bdev_raid.raid1_resize_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:17.300 15:14:07 bdev_raid.raid1_resize_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:17.300 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:17.300 15:14:07 bdev_raid.raid1_resize_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:17.300 15:14:07 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.300 [2024-11-19 15:14:07.610601] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:08:17.300 [2024-11-19 15:14:07.610832] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:17.560 [2024-11-19 15:14:07.769557] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:17.560 [2024-11-19 15:14:07.807795] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:17.560 [2024-11-19 15:14:07.884268] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:17.560 [2024-11-19 15:14:07.884312] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:18.130 15:14:08 bdev_raid.raid1_resize_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:18.130 15:14:08 bdev_raid.raid1_resize_test -- common/autotest_common.sh@868 -- # return 0 00:08:18.130 15:14:08 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@346 -- # rpc_cmd bdev_null_create Base_1 32 512 00:08:18.130 15:14:08 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:18.130 15:14:08 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.130 Base_1 00:08:18.130 15:14:08 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:18.130 15:14:08 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@347 -- # rpc_cmd bdev_null_create Base_2 32 512 00:08:18.130 15:14:08 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:18.130 15:14:08 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.130 Base_2 00:08:18.130 15:14:08 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:18.130 15:14:08 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@349 -- # '[' 1 -eq 0 ']' 00:08:18.130 15:14:08 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@352 -- # rpc_cmd bdev_raid_create -r 1 -b ''\''Base_1 Base_2'\''' -n Raid 00:08:18.130 15:14:08 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:18.130 15:14:08 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.130 [2024-11-19 15:14:08.455614] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:08:18.130 [2024-11-19 15:14:08.457785] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:08:18.130 [2024-11-19 15:14:08.457934] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:08:18.130 [2024-11-19 15:14:08.457951] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:08:18.130 [2024-11-19 15:14:08.458273] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000021f0 00:08:18.130 [2024-11-19 15:14:08.458382] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:08:18.130 [2024-11-19 15:14:08.458390] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000001200 00:08:18.130 [2024-11-19 15:14:08.458497] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:18.130 15:14:08 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:18.130 15:14:08 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@356 -- # rpc_cmd bdev_null_resize Base_1 64 00:08:18.130 15:14:08 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:18.130 15:14:08 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.130 [2024-11-19 15:14:08.463572] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:08:18.130 [2024-11-19 15:14:08.463605] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_1' was resized: old size 65536, new size 131072 00:08:18.390 true 00:08:18.390 15:14:08 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:18.390 15:14:08 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@359 -- # rpc_cmd bdev_get_bdevs -b Raid 00:08:18.390 15:14:08 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@359 -- # jq '.[].num_blocks' 00:08:18.390 15:14:08 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:18.391 15:14:08 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.391 [2024-11-19 15:14:08.479720] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:18.391 15:14:08 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:18.391 15:14:08 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@359 -- # blkcnt=65536 00:08:18.391 15:14:08 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@360 -- # raid_size_mb=32 00:08:18.391 15:14:08 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@361 -- # '[' 1 -eq 0 ']' 00:08:18.391 15:14:08 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@364 -- # expected_size=32 00:08:18.391 15:14:08 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@366 -- # '[' 32 '!=' 32 ']' 00:08:18.391 15:14:08 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@372 -- # rpc_cmd bdev_null_resize Base_2 64 00:08:18.391 15:14:08 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:18.391 15:14:08 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.391 [2024-11-19 15:14:08.523429] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:08:18.391 [2024-11-19 15:14:08.523493] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_2' was resized: old size 65536, new size 131072 00:08:18.391 [2024-11-19 15:14:08.523545] bdev_raid.c:2344:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 65536 to 131072 00:08:18.391 true 00:08:18.391 15:14:08 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:18.391 15:14:08 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@375 -- # rpc_cmd bdev_get_bdevs -b Raid 00:08:18.391 15:14:08 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@375 -- # jq '.[].num_blocks' 00:08:18.391 15:14:08 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:18.391 15:14:08 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.391 [2024-11-19 15:14:08.539586] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:18.391 15:14:08 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:18.391 15:14:08 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@375 -- # blkcnt=131072 00:08:18.391 15:14:08 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@376 -- # raid_size_mb=64 00:08:18.391 15:14:08 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@377 -- # '[' 1 -eq 0 ']' 00:08:18.391 15:14:08 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@380 -- # expected_size=64 00:08:18.391 15:14:08 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@382 -- # '[' 64 '!=' 64 ']' 00:08:18.391 15:14:08 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@387 -- # killprocess 72035 00:08:18.391 15:14:08 bdev_raid.raid1_resize_test -- common/autotest_common.sh@954 -- # '[' -z 72035 ']' 00:08:18.391 15:14:08 bdev_raid.raid1_resize_test -- common/autotest_common.sh@958 -- # kill -0 72035 00:08:18.391 15:14:08 bdev_raid.raid1_resize_test -- common/autotest_common.sh@959 -- # uname 00:08:18.391 15:14:08 bdev_raid.raid1_resize_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:18.391 15:14:08 bdev_raid.raid1_resize_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 72035 00:08:18.391 killing process with pid 72035 00:08:18.391 15:14:08 bdev_raid.raid1_resize_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:18.391 15:14:08 bdev_raid.raid1_resize_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:18.391 15:14:08 bdev_raid.raid1_resize_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 72035' 00:08:18.391 15:14:08 bdev_raid.raid1_resize_test -- common/autotest_common.sh@973 -- # kill 72035 00:08:18.391 [2024-11-19 15:14:08.624563] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:18.391 [2024-11-19 15:14:08.624633] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:18.391 15:14:08 bdev_raid.raid1_resize_test -- common/autotest_common.sh@978 -- # wait 72035 00:08:18.391 [2024-11-19 15:14:08.625093] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:18.391 [2024-11-19 15:14:08.625164] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Raid, state offline 00:08:18.391 [2024-11-19 15:14:08.626868] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:18.651 15:14:08 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@389 -- # return 0 00:08:18.651 00:08:18.651 real 0m1.426s 00:08:18.651 user 0m1.513s 00:08:18.651 sys 0m0.384s 00:08:18.651 15:14:08 bdev_raid.raid1_resize_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:18.651 15:14:08 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.651 ************************************ 00:08:18.651 END TEST raid1_resize_test 00:08:18.651 ************************************ 00:08:18.910 15:14:09 bdev_raid -- bdev/bdev_raid.sh@966 -- # for n in {2..4} 00:08:18.910 15:14:09 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:08:18.910 15:14:09 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid0 2 false 00:08:18.910 15:14:09 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:18.910 15:14:09 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:18.910 15:14:09 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:18.910 ************************************ 00:08:18.910 START TEST raid_state_function_test 00:08:18.910 ************************************ 00:08:18.910 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid0 2 false 00:08:18.910 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:08:18.910 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:08:18.910 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:08:18.910 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:18.910 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:18.910 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:18.910 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:18.910 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:18.910 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:18.910 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:18.910 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:18.910 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:18.910 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:08:18.910 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:18.910 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:18.911 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:18.911 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:18.911 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:18.911 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:08:18.911 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:08:18.911 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:08:18.911 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:08:18.911 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:08:18.911 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=72086 00:08:18.911 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:18.911 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 72086' 00:08:18.911 Process raid pid: 72086 00:08:18.911 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 72086 00:08:18.911 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 72086 ']' 00:08:18.911 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:18.911 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:18.911 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:18.911 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:18.911 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:18.911 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.911 [2024-11-19 15:14:09.110395] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:08:18.911 [2024-11-19 15:14:09.110615] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:19.169 [2024-11-19 15:14:09.265873] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:19.169 [2024-11-19 15:14:09.304408] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:19.169 [2024-11-19 15:14:09.380263] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:19.169 [2024-11-19 15:14:09.380304] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:19.738 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:19.738 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:08:19.738 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:08:19.738 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:19.738 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.738 [2024-11-19 15:14:09.951655] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:19.738 [2024-11-19 15:14:09.951847] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:19.738 [2024-11-19 15:14:09.951863] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:19.738 [2024-11-19 15:14:09.951875] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:19.738 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:19.738 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:08:19.738 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:19.738 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:19.738 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:19.738 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:19.738 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:19.738 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:19.738 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:19.738 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:19.738 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:19.738 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:19.738 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:19.738 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:19.738 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.738 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:19.738 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:19.738 "name": "Existed_Raid", 00:08:19.738 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:19.738 "strip_size_kb": 64, 00:08:19.738 "state": "configuring", 00:08:19.738 "raid_level": "raid0", 00:08:19.738 "superblock": false, 00:08:19.738 "num_base_bdevs": 2, 00:08:19.738 "num_base_bdevs_discovered": 0, 00:08:19.738 "num_base_bdevs_operational": 2, 00:08:19.738 "base_bdevs_list": [ 00:08:19.738 { 00:08:19.738 "name": "BaseBdev1", 00:08:19.738 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:19.738 "is_configured": false, 00:08:19.738 "data_offset": 0, 00:08:19.738 "data_size": 0 00:08:19.738 }, 00:08:19.738 { 00:08:19.738 "name": "BaseBdev2", 00:08:19.738 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:19.738 "is_configured": false, 00:08:19.738 "data_offset": 0, 00:08:19.738 "data_size": 0 00:08:19.738 } 00:08:19.738 ] 00:08:19.738 }' 00:08:19.738 15:14:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:19.738 15:14:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.305 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:20.305 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:20.305 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.305 [2024-11-19 15:14:10.366857] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:20.306 [2024-11-19 15:14:10.367008] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:08:20.306 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:20.306 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:08:20.306 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:20.306 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.306 [2024-11-19 15:14:10.378807] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:20.306 [2024-11-19 15:14:10.378902] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:20.306 [2024-11-19 15:14:10.378942] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:20.306 [2024-11-19 15:14:10.379015] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:20.306 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:20.306 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:20.306 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:20.306 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.306 [2024-11-19 15:14:10.405951] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:20.306 BaseBdev1 00:08:20.306 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:20.306 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:08:20.306 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:08:20.306 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:20.306 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:20.306 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:20.306 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:20.306 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:20.306 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:20.306 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.306 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:20.306 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:20.306 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:20.306 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.306 [ 00:08:20.306 { 00:08:20.306 "name": "BaseBdev1", 00:08:20.306 "aliases": [ 00:08:20.306 "d4a88657-2874-4380-badc-b301277553e8" 00:08:20.306 ], 00:08:20.306 "product_name": "Malloc disk", 00:08:20.306 "block_size": 512, 00:08:20.306 "num_blocks": 65536, 00:08:20.306 "uuid": "d4a88657-2874-4380-badc-b301277553e8", 00:08:20.306 "assigned_rate_limits": { 00:08:20.306 "rw_ios_per_sec": 0, 00:08:20.306 "rw_mbytes_per_sec": 0, 00:08:20.306 "r_mbytes_per_sec": 0, 00:08:20.306 "w_mbytes_per_sec": 0 00:08:20.306 }, 00:08:20.306 "claimed": true, 00:08:20.306 "claim_type": "exclusive_write", 00:08:20.306 "zoned": false, 00:08:20.306 "supported_io_types": { 00:08:20.306 "read": true, 00:08:20.306 "write": true, 00:08:20.306 "unmap": true, 00:08:20.306 "flush": true, 00:08:20.306 "reset": true, 00:08:20.306 "nvme_admin": false, 00:08:20.306 "nvme_io": false, 00:08:20.306 "nvme_io_md": false, 00:08:20.306 "write_zeroes": true, 00:08:20.306 "zcopy": true, 00:08:20.306 "get_zone_info": false, 00:08:20.306 "zone_management": false, 00:08:20.306 "zone_append": false, 00:08:20.306 "compare": false, 00:08:20.306 "compare_and_write": false, 00:08:20.306 "abort": true, 00:08:20.306 "seek_hole": false, 00:08:20.306 "seek_data": false, 00:08:20.306 "copy": true, 00:08:20.306 "nvme_iov_md": false 00:08:20.306 }, 00:08:20.306 "memory_domains": [ 00:08:20.306 { 00:08:20.306 "dma_device_id": "system", 00:08:20.306 "dma_device_type": 1 00:08:20.306 }, 00:08:20.306 { 00:08:20.306 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:20.306 "dma_device_type": 2 00:08:20.306 } 00:08:20.306 ], 00:08:20.306 "driver_specific": {} 00:08:20.306 } 00:08:20.306 ] 00:08:20.306 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:20.306 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:20.306 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:08:20.306 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:20.306 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:20.306 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:20.306 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:20.306 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:20.306 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:20.306 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:20.306 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:20.306 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:20.306 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:20.306 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:20.306 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.306 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:20.306 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:20.306 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:20.306 "name": "Existed_Raid", 00:08:20.306 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:20.306 "strip_size_kb": 64, 00:08:20.306 "state": "configuring", 00:08:20.306 "raid_level": "raid0", 00:08:20.306 "superblock": false, 00:08:20.306 "num_base_bdevs": 2, 00:08:20.306 "num_base_bdevs_discovered": 1, 00:08:20.306 "num_base_bdevs_operational": 2, 00:08:20.306 "base_bdevs_list": [ 00:08:20.306 { 00:08:20.306 "name": "BaseBdev1", 00:08:20.306 "uuid": "d4a88657-2874-4380-badc-b301277553e8", 00:08:20.306 "is_configured": true, 00:08:20.306 "data_offset": 0, 00:08:20.306 "data_size": 65536 00:08:20.306 }, 00:08:20.306 { 00:08:20.306 "name": "BaseBdev2", 00:08:20.306 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:20.306 "is_configured": false, 00:08:20.306 "data_offset": 0, 00:08:20.306 "data_size": 0 00:08:20.306 } 00:08:20.306 ] 00:08:20.306 }' 00:08:20.306 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:20.306 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.565 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:20.565 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:20.565 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.565 [2024-11-19 15:14:10.853257] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:20.565 [2024-11-19 15:14:10.853341] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:08:20.565 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:20.565 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:08:20.565 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:20.565 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.565 [2024-11-19 15:14:10.865226] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:20.565 [2024-11-19 15:14:10.867447] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:20.565 [2024-11-19 15:14:10.867540] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:20.565 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:20.565 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:08:20.565 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:20.565 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:08:20.565 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:20.565 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:20.565 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:20.565 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:20.566 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:20.566 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:20.566 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:20.566 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:20.566 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:20.566 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:20.566 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:20.566 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:20.566 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.566 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:20.825 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:20.825 "name": "Existed_Raid", 00:08:20.825 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:20.825 "strip_size_kb": 64, 00:08:20.825 "state": "configuring", 00:08:20.825 "raid_level": "raid0", 00:08:20.825 "superblock": false, 00:08:20.825 "num_base_bdevs": 2, 00:08:20.825 "num_base_bdevs_discovered": 1, 00:08:20.825 "num_base_bdevs_operational": 2, 00:08:20.825 "base_bdevs_list": [ 00:08:20.825 { 00:08:20.825 "name": "BaseBdev1", 00:08:20.825 "uuid": "d4a88657-2874-4380-badc-b301277553e8", 00:08:20.825 "is_configured": true, 00:08:20.825 "data_offset": 0, 00:08:20.825 "data_size": 65536 00:08:20.825 }, 00:08:20.825 { 00:08:20.825 "name": "BaseBdev2", 00:08:20.825 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:20.825 "is_configured": false, 00:08:20.825 "data_offset": 0, 00:08:20.825 "data_size": 0 00:08:20.825 } 00:08:20.825 ] 00:08:20.825 }' 00:08:20.825 15:14:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:20.825 15:14:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.084 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:21.084 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:21.084 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.084 [2024-11-19 15:14:11.281273] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:21.084 [2024-11-19 15:14:11.281396] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:21.084 [2024-11-19 15:14:11.281419] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:08:21.084 [2024-11-19 15:14:11.281731] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:08:21.084 [2024-11-19 15:14:11.281905] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:21.084 [2024-11-19 15:14:11.281920] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:08:21.084 [2024-11-19 15:14:11.282157] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:21.084 BaseBdev2 00:08:21.084 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:21.084 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:08:21.084 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:08:21.084 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:21.084 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:21.084 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:21.084 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:21.084 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:21.084 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:21.084 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.084 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:21.084 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:21.084 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:21.084 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.084 [ 00:08:21.084 { 00:08:21.084 "name": "BaseBdev2", 00:08:21.084 "aliases": [ 00:08:21.084 "6750917e-a285-4ea3-9774-0bc31ab06d04" 00:08:21.084 ], 00:08:21.084 "product_name": "Malloc disk", 00:08:21.084 "block_size": 512, 00:08:21.084 "num_blocks": 65536, 00:08:21.084 "uuid": "6750917e-a285-4ea3-9774-0bc31ab06d04", 00:08:21.084 "assigned_rate_limits": { 00:08:21.084 "rw_ios_per_sec": 0, 00:08:21.084 "rw_mbytes_per_sec": 0, 00:08:21.084 "r_mbytes_per_sec": 0, 00:08:21.084 "w_mbytes_per_sec": 0 00:08:21.084 }, 00:08:21.084 "claimed": true, 00:08:21.084 "claim_type": "exclusive_write", 00:08:21.084 "zoned": false, 00:08:21.084 "supported_io_types": { 00:08:21.084 "read": true, 00:08:21.084 "write": true, 00:08:21.084 "unmap": true, 00:08:21.084 "flush": true, 00:08:21.084 "reset": true, 00:08:21.084 "nvme_admin": false, 00:08:21.084 "nvme_io": false, 00:08:21.084 "nvme_io_md": false, 00:08:21.084 "write_zeroes": true, 00:08:21.084 "zcopy": true, 00:08:21.084 "get_zone_info": false, 00:08:21.084 "zone_management": false, 00:08:21.084 "zone_append": false, 00:08:21.084 "compare": false, 00:08:21.084 "compare_and_write": false, 00:08:21.084 "abort": true, 00:08:21.084 "seek_hole": false, 00:08:21.084 "seek_data": false, 00:08:21.084 "copy": true, 00:08:21.084 "nvme_iov_md": false 00:08:21.084 }, 00:08:21.084 "memory_domains": [ 00:08:21.084 { 00:08:21.084 "dma_device_id": "system", 00:08:21.084 "dma_device_type": 1 00:08:21.084 }, 00:08:21.084 { 00:08:21.084 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:21.084 "dma_device_type": 2 00:08:21.084 } 00:08:21.084 ], 00:08:21.084 "driver_specific": {} 00:08:21.084 } 00:08:21.084 ] 00:08:21.084 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:21.084 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:21.084 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:21.084 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:21.084 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 2 00:08:21.084 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:21.084 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:21.084 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:21.084 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:21.084 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:21.084 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:21.084 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:21.084 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:21.084 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:21.084 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:21.084 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:21.084 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:21.084 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.084 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:21.084 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:21.084 "name": "Existed_Raid", 00:08:21.084 "uuid": "fe642823-f54a-42b2-9241-48f21a455034", 00:08:21.084 "strip_size_kb": 64, 00:08:21.084 "state": "online", 00:08:21.084 "raid_level": "raid0", 00:08:21.084 "superblock": false, 00:08:21.084 "num_base_bdevs": 2, 00:08:21.084 "num_base_bdevs_discovered": 2, 00:08:21.084 "num_base_bdevs_operational": 2, 00:08:21.084 "base_bdevs_list": [ 00:08:21.084 { 00:08:21.084 "name": "BaseBdev1", 00:08:21.084 "uuid": "d4a88657-2874-4380-badc-b301277553e8", 00:08:21.084 "is_configured": true, 00:08:21.084 "data_offset": 0, 00:08:21.084 "data_size": 65536 00:08:21.084 }, 00:08:21.084 { 00:08:21.084 "name": "BaseBdev2", 00:08:21.084 "uuid": "6750917e-a285-4ea3-9774-0bc31ab06d04", 00:08:21.084 "is_configured": true, 00:08:21.084 "data_offset": 0, 00:08:21.084 "data_size": 65536 00:08:21.084 } 00:08:21.084 ] 00:08:21.084 }' 00:08:21.084 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:21.084 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.654 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:08:21.654 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:21.654 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:21.654 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:21.654 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:21.654 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:21.654 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:21.654 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:21.654 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:21.654 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.654 [2024-11-19 15:14:11.736888] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:21.654 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:21.654 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:21.654 "name": "Existed_Raid", 00:08:21.654 "aliases": [ 00:08:21.654 "fe642823-f54a-42b2-9241-48f21a455034" 00:08:21.654 ], 00:08:21.654 "product_name": "Raid Volume", 00:08:21.654 "block_size": 512, 00:08:21.654 "num_blocks": 131072, 00:08:21.654 "uuid": "fe642823-f54a-42b2-9241-48f21a455034", 00:08:21.654 "assigned_rate_limits": { 00:08:21.654 "rw_ios_per_sec": 0, 00:08:21.654 "rw_mbytes_per_sec": 0, 00:08:21.654 "r_mbytes_per_sec": 0, 00:08:21.654 "w_mbytes_per_sec": 0 00:08:21.654 }, 00:08:21.654 "claimed": false, 00:08:21.654 "zoned": false, 00:08:21.654 "supported_io_types": { 00:08:21.654 "read": true, 00:08:21.654 "write": true, 00:08:21.654 "unmap": true, 00:08:21.654 "flush": true, 00:08:21.654 "reset": true, 00:08:21.654 "nvme_admin": false, 00:08:21.654 "nvme_io": false, 00:08:21.654 "nvme_io_md": false, 00:08:21.654 "write_zeroes": true, 00:08:21.654 "zcopy": false, 00:08:21.654 "get_zone_info": false, 00:08:21.654 "zone_management": false, 00:08:21.654 "zone_append": false, 00:08:21.654 "compare": false, 00:08:21.654 "compare_and_write": false, 00:08:21.654 "abort": false, 00:08:21.654 "seek_hole": false, 00:08:21.654 "seek_data": false, 00:08:21.654 "copy": false, 00:08:21.654 "nvme_iov_md": false 00:08:21.654 }, 00:08:21.654 "memory_domains": [ 00:08:21.654 { 00:08:21.654 "dma_device_id": "system", 00:08:21.654 "dma_device_type": 1 00:08:21.654 }, 00:08:21.654 { 00:08:21.654 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:21.654 "dma_device_type": 2 00:08:21.654 }, 00:08:21.654 { 00:08:21.654 "dma_device_id": "system", 00:08:21.654 "dma_device_type": 1 00:08:21.654 }, 00:08:21.654 { 00:08:21.654 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:21.654 "dma_device_type": 2 00:08:21.654 } 00:08:21.654 ], 00:08:21.654 "driver_specific": { 00:08:21.654 "raid": { 00:08:21.654 "uuid": "fe642823-f54a-42b2-9241-48f21a455034", 00:08:21.654 "strip_size_kb": 64, 00:08:21.654 "state": "online", 00:08:21.654 "raid_level": "raid0", 00:08:21.654 "superblock": false, 00:08:21.654 "num_base_bdevs": 2, 00:08:21.654 "num_base_bdevs_discovered": 2, 00:08:21.654 "num_base_bdevs_operational": 2, 00:08:21.654 "base_bdevs_list": [ 00:08:21.654 { 00:08:21.654 "name": "BaseBdev1", 00:08:21.654 "uuid": "d4a88657-2874-4380-badc-b301277553e8", 00:08:21.654 "is_configured": true, 00:08:21.655 "data_offset": 0, 00:08:21.655 "data_size": 65536 00:08:21.655 }, 00:08:21.655 { 00:08:21.655 "name": "BaseBdev2", 00:08:21.655 "uuid": "6750917e-a285-4ea3-9774-0bc31ab06d04", 00:08:21.655 "is_configured": true, 00:08:21.655 "data_offset": 0, 00:08:21.655 "data_size": 65536 00:08:21.655 } 00:08:21.655 ] 00:08:21.655 } 00:08:21.655 } 00:08:21.655 }' 00:08:21.655 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:21.655 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:08:21.655 BaseBdev2' 00:08:21.655 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:21.655 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:21.655 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:21.655 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:08:21.655 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:21.655 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.655 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:21.655 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:21.655 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:21.655 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:21.655 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:21.655 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:21.655 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:21.655 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.655 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:21.655 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:21.655 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:21.655 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:21.655 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:21.655 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:21.655 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.655 [2024-11-19 15:14:11.928386] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:21.655 [2024-11-19 15:14:11.928436] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:21.655 [2024-11-19 15:14:11.928497] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:21.655 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:21.655 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:08:21.655 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:08:21.655 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:21.655 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:21.655 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:08:21.655 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 1 00:08:21.655 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:21.655 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:08:21.655 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:21.655 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:21.655 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:08:21.655 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:21.655 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:21.655 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:21.655 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:21.655 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:21.655 15:14:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:21.655 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:21.655 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.655 15:14:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:21.915 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:21.915 "name": "Existed_Raid", 00:08:21.915 "uuid": "fe642823-f54a-42b2-9241-48f21a455034", 00:08:21.915 "strip_size_kb": 64, 00:08:21.915 "state": "offline", 00:08:21.915 "raid_level": "raid0", 00:08:21.915 "superblock": false, 00:08:21.915 "num_base_bdevs": 2, 00:08:21.915 "num_base_bdevs_discovered": 1, 00:08:21.915 "num_base_bdevs_operational": 1, 00:08:21.915 "base_bdevs_list": [ 00:08:21.915 { 00:08:21.915 "name": null, 00:08:21.915 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:21.915 "is_configured": false, 00:08:21.915 "data_offset": 0, 00:08:21.915 "data_size": 65536 00:08:21.915 }, 00:08:21.915 { 00:08:21.915 "name": "BaseBdev2", 00:08:21.915 "uuid": "6750917e-a285-4ea3-9774-0bc31ab06d04", 00:08:21.915 "is_configured": true, 00:08:21.915 "data_offset": 0, 00:08:21.915 "data_size": 65536 00:08:21.915 } 00:08:21.915 ] 00:08:21.915 }' 00:08:21.915 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:21.915 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.175 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:08:22.175 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:22.175 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:22.175 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:22.175 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.175 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:22.175 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:22.175 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:22.175 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:22.175 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:08:22.175 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:22.175 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.175 [2024-11-19 15:14:12.416505] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:22.175 [2024-11-19 15:14:12.416652] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:08:22.175 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:22.175 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:22.175 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:22.175 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:22.175 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:22.175 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.175 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:08:22.175 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:22.175 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:08:22.175 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:08:22.175 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:08:22.175 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 72086 00:08:22.175 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 72086 ']' 00:08:22.175 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 72086 00:08:22.175 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:08:22.175 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:22.175 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 72086 00:08:22.175 killing process with pid 72086 00:08:22.175 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:22.175 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:22.175 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 72086' 00:08:22.175 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 72086 00:08:22.175 [2024-11-19 15:14:12.512568] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:22.175 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 72086 00:08:22.435 [2024-11-19 15:14:12.514180] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:22.695 15:14:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:08:22.695 00:08:22.695 real 0m3.813s 00:08:22.695 user 0m5.832s 00:08:22.695 sys 0m0.774s 00:08:22.695 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:22.695 15:14:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.695 ************************************ 00:08:22.695 END TEST raid_state_function_test 00:08:22.695 ************************************ 00:08:22.695 15:14:12 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid0 2 true 00:08:22.695 15:14:12 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:22.695 15:14:12 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:22.695 15:14:12 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:22.695 ************************************ 00:08:22.695 START TEST raid_state_function_test_sb 00:08:22.695 ************************************ 00:08:22.695 15:14:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid0 2 true 00:08:22.695 15:14:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:08:22.695 15:14:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:08:22.695 15:14:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:08:22.695 15:14:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:22.695 15:14:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:22.695 15:14:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:22.695 15:14:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:22.695 15:14:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:22.695 15:14:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:22.695 15:14:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:22.695 15:14:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:22.695 15:14:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:22.695 15:14:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:08:22.695 15:14:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:22.695 15:14:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:22.695 15:14:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:22.695 15:14:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:22.695 15:14:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:22.695 15:14:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:08:22.695 15:14:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:08:22.695 15:14:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:08:22.695 15:14:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:08:22.695 15:14:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:08:22.695 15:14:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=72323 00:08:22.695 Process raid pid: 72323 00:08:22.695 15:14:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:22.695 15:14:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 72323' 00:08:22.695 15:14:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 72323 00:08:22.695 15:14:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 72323 ']' 00:08:22.695 15:14:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:22.695 15:14:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:22.695 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:22.695 15:14:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:22.695 15:14:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:22.695 15:14:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:22.695 [2024-11-19 15:14:12.996861] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:08:22.696 [2024-11-19 15:14:12.997000] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:22.957 [2024-11-19 15:14:13.151226] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:22.957 [2024-11-19 15:14:13.189405] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:22.957 [2024-11-19 15:14:13.267071] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:22.957 [2024-11-19 15:14:13.267172] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:23.527 15:14:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:23.527 15:14:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:08:23.527 15:14:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:08:23.527 15:14:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.527 15:14:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:23.527 [2024-11-19 15:14:13.858563] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:23.527 [2024-11-19 15:14:13.858657] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:23.527 [2024-11-19 15:14:13.858668] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:23.527 [2024-11-19 15:14:13.858678] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:23.787 15:14:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.787 15:14:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:08:23.787 15:14:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:23.787 15:14:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:23.787 15:14:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:23.787 15:14:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:23.787 15:14:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:23.787 15:14:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:23.787 15:14:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:23.787 15:14:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:23.787 15:14:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:23.787 15:14:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:23.787 15:14:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:23.787 15:14:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.787 15:14:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:23.787 15:14:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.787 15:14:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:23.787 "name": "Existed_Raid", 00:08:23.787 "uuid": "4b151186-f78a-4eb9-bb09-3d35eb0e841b", 00:08:23.787 "strip_size_kb": 64, 00:08:23.787 "state": "configuring", 00:08:23.787 "raid_level": "raid0", 00:08:23.787 "superblock": true, 00:08:23.787 "num_base_bdevs": 2, 00:08:23.787 "num_base_bdevs_discovered": 0, 00:08:23.787 "num_base_bdevs_operational": 2, 00:08:23.787 "base_bdevs_list": [ 00:08:23.787 { 00:08:23.787 "name": "BaseBdev1", 00:08:23.787 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:23.787 "is_configured": false, 00:08:23.787 "data_offset": 0, 00:08:23.787 "data_size": 0 00:08:23.787 }, 00:08:23.787 { 00:08:23.787 "name": "BaseBdev2", 00:08:23.787 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:23.787 "is_configured": false, 00:08:23.787 "data_offset": 0, 00:08:23.787 "data_size": 0 00:08:23.787 } 00:08:23.787 ] 00:08:23.787 }' 00:08:23.787 15:14:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:23.787 15:14:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:24.048 15:14:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:24.048 15:14:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:24.048 15:14:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:24.048 [2024-11-19 15:14:14.281754] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:24.048 [2024-11-19 15:14:14.281888] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:08:24.048 15:14:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:24.048 15:14:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:08:24.048 15:14:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:24.048 15:14:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:24.048 [2024-11-19 15:14:14.293713] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:24.048 [2024-11-19 15:14:14.293823] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:24.048 [2024-11-19 15:14:14.293851] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:24.048 [2024-11-19 15:14:14.293888] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:24.048 15:14:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:24.048 15:14:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:24.048 15:14:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:24.048 15:14:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:24.048 [2024-11-19 15:14:14.320843] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:24.048 BaseBdev1 00:08:24.048 15:14:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:24.048 15:14:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:08:24.048 15:14:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:08:24.048 15:14:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:24.048 15:14:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:24.048 15:14:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:24.048 15:14:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:24.048 15:14:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:24.048 15:14:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:24.048 15:14:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:24.048 15:14:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:24.048 15:14:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:24.048 15:14:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:24.048 15:14:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:24.048 [ 00:08:24.048 { 00:08:24.048 "name": "BaseBdev1", 00:08:24.048 "aliases": [ 00:08:24.048 "b2a14192-e58a-4d64-99c5-e8ecad417653" 00:08:24.048 ], 00:08:24.048 "product_name": "Malloc disk", 00:08:24.048 "block_size": 512, 00:08:24.048 "num_blocks": 65536, 00:08:24.048 "uuid": "b2a14192-e58a-4d64-99c5-e8ecad417653", 00:08:24.048 "assigned_rate_limits": { 00:08:24.048 "rw_ios_per_sec": 0, 00:08:24.048 "rw_mbytes_per_sec": 0, 00:08:24.048 "r_mbytes_per_sec": 0, 00:08:24.048 "w_mbytes_per_sec": 0 00:08:24.048 }, 00:08:24.048 "claimed": true, 00:08:24.048 "claim_type": "exclusive_write", 00:08:24.048 "zoned": false, 00:08:24.048 "supported_io_types": { 00:08:24.048 "read": true, 00:08:24.048 "write": true, 00:08:24.048 "unmap": true, 00:08:24.048 "flush": true, 00:08:24.048 "reset": true, 00:08:24.048 "nvme_admin": false, 00:08:24.048 "nvme_io": false, 00:08:24.048 "nvme_io_md": false, 00:08:24.048 "write_zeroes": true, 00:08:24.048 "zcopy": true, 00:08:24.048 "get_zone_info": false, 00:08:24.048 "zone_management": false, 00:08:24.048 "zone_append": false, 00:08:24.048 "compare": false, 00:08:24.048 "compare_and_write": false, 00:08:24.048 "abort": true, 00:08:24.048 "seek_hole": false, 00:08:24.048 "seek_data": false, 00:08:24.048 "copy": true, 00:08:24.048 "nvme_iov_md": false 00:08:24.048 }, 00:08:24.048 "memory_domains": [ 00:08:24.048 { 00:08:24.048 "dma_device_id": "system", 00:08:24.048 "dma_device_type": 1 00:08:24.048 }, 00:08:24.048 { 00:08:24.048 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:24.048 "dma_device_type": 2 00:08:24.048 } 00:08:24.048 ], 00:08:24.048 "driver_specific": {} 00:08:24.048 } 00:08:24.048 ] 00:08:24.048 15:14:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:24.048 15:14:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:24.048 15:14:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:08:24.048 15:14:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:24.048 15:14:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:24.048 15:14:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:24.048 15:14:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:24.048 15:14:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:24.048 15:14:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:24.048 15:14:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:24.048 15:14:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:24.048 15:14:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:24.048 15:14:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:24.048 15:14:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:24.048 15:14:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:24.048 15:14:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:24.048 15:14:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:24.309 15:14:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:24.309 "name": "Existed_Raid", 00:08:24.309 "uuid": "f4301d46-06c4-4e15-aa3a-a7e956e25a34", 00:08:24.309 "strip_size_kb": 64, 00:08:24.309 "state": "configuring", 00:08:24.309 "raid_level": "raid0", 00:08:24.309 "superblock": true, 00:08:24.309 "num_base_bdevs": 2, 00:08:24.309 "num_base_bdevs_discovered": 1, 00:08:24.309 "num_base_bdevs_operational": 2, 00:08:24.309 "base_bdevs_list": [ 00:08:24.309 { 00:08:24.309 "name": "BaseBdev1", 00:08:24.309 "uuid": "b2a14192-e58a-4d64-99c5-e8ecad417653", 00:08:24.309 "is_configured": true, 00:08:24.309 "data_offset": 2048, 00:08:24.309 "data_size": 63488 00:08:24.309 }, 00:08:24.309 { 00:08:24.309 "name": "BaseBdev2", 00:08:24.309 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:24.309 "is_configured": false, 00:08:24.309 "data_offset": 0, 00:08:24.309 "data_size": 0 00:08:24.309 } 00:08:24.309 ] 00:08:24.309 }' 00:08:24.309 15:14:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:24.309 15:14:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:24.569 15:14:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:24.569 15:14:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:24.569 15:14:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:24.569 [2024-11-19 15:14:14.800110] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:24.569 [2024-11-19 15:14:14.800183] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:08:24.569 15:14:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:24.569 15:14:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:08:24.569 15:14:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:24.569 15:14:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:24.569 [2024-11-19 15:14:14.808105] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:24.569 [2024-11-19 15:14:14.810318] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:24.569 [2024-11-19 15:14:14.810437] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:24.569 15:14:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:24.569 15:14:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:08:24.569 15:14:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:24.569 15:14:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:08:24.569 15:14:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:24.569 15:14:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:24.569 15:14:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:24.570 15:14:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:24.570 15:14:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:24.570 15:14:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:24.570 15:14:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:24.570 15:14:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:24.570 15:14:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:24.570 15:14:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:24.570 15:14:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:24.570 15:14:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:24.570 15:14:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:24.570 15:14:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:24.570 15:14:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:24.570 "name": "Existed_Raid", 00:08:24.570 "uuid": "6d7b0fe5-9c30-4c7d-b749-584a444e5c35", 00:08:24.570 "strip_size_kb": 64, 00:08:24.570 "state": "configuring", 00:08:24.570 "raid_level": "raid0", 00:08:24.570 "superblock": true, 00:08:24.570 "num_base_bdevs": 2, 00:08:24.570 "num_base_bdevs_discovered": 1, 00:08:24.570 "num_base_bdevs_operational": 2, 00:08:24.570 "base_bdevs_list": [ 00:08:24.570 { 00:08:24.570 "name": "BaseBdev1", 00:08:24.570 "uuid": "b2a14192-e58a-4d64-99c5-e8ecad417653", 00:08:24.570 "is_configured": true, 00:08:24.570 "data_offset": 2048, 00:08:24.570 "data_size": 63488 00:08:24.570 }, 00:08:24.570 { 00:08:24.570 "name": "BaseBdev2", 00:08:24.570 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:24.570 "is_configured": false, 00:08:24.570 "data_offset": 0, 00:08:24.570 "data_size": 0 00:08:24.570 } 00:08:24.570 ] 00:08:24.570 }' 00:08:24.570 15:14:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:24.570 15:14:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.139 15:14:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:25.139 15:14:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:25.139 15:14:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.139 [2024-11-19 15:14:15.288123] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:25.139 [2024-11-19 15:14:15.288454] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:25.139 [2024-11-19 15:14:15.288497] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:08:25.139 BaseBdev2 00:08:25.139 [2024-11-19 15:14:15.288880] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:08:25.139 [2024-11-19 15:14:15.289102] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:25.139 [2024-11-19 15:14:15.289153] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:08:25.139 [2024-11-19 15:14:15.289317] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:25.139 15:14:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:25.139 15:14:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:08:25.139 15:14:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:08:25.139 15:14:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:25.139 15:14:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:25.139 15:14:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:25.139 15:14:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:25.139 15:14:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:25.139 15:14:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:25.139 15:14:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.139 15:14:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:25.139 15:14:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:25.140 15:14:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:25.140 15:14:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.140 [ 00:08:25.140 { 00:08:25.140 "name": "BaseBdev2", 00:08:25.140 "aliases": [ 00:08:25.140 "0932504a-7dc3-453d-973c-bc2504031e1e" 00:08:25.140 ], 00:08:25.140 "product_name": "Malloc disk", 00:08:25.140 "block_size": 512, 00:08:25.140 "num_blocks": 65536, 00:08:25.140 "uuid": "0932504a-7dc3-453d-973c-bc2504031e1e", 00:08:25.140 "assigned_rate_limits": { 00:08:25.140 "rw_ios_per_sec": 0, 00:08:25.140 "rw_mbytes_per_sec": 0, 00:08:25.140 "r_mbytes_per_sec": 0, 00:08:25.140 "w_mbytes_per_sec": 0 00:08:25.140 }, 00:08:25.140 "claimed": true, 00:08:25.140 "claim_type": "exclusive_write", 00:08:25.140 "zoned": false, 00:08:25.140 "supported_io_types": { 00:08:25.140 "read": true, 00:08:25.140 "write": true, 00:08:25.140 "unmap": true, 00:08:25.140 "flush": true, 00:08:25.140 "reset": true, 00:08:25.140 "nvme_admin": false, 00:08:25.140 "nvme_io": false, 00:08:25.140 "nvme_io_md": false, 00:08:25.140 "write_zeroes": true, 00:08:25.140 "zcopy": true, 00:08:25.140 "get_zone_info": false, 00:08:25.140 "zone_management": false, 00:08:25.140 "zone_append": false, 00:08:25.140 "compare": false, 00:08:25.140 "compare_and_write": false, 00:08:25.140 "abort": true, 00:08:25.140 "seek_hole": false, 00:08:25.140 "seek_data": false, 00:08:25.140 "copy": true, 00:08:25.140 "nvme_iov_md": false 00:08:25.140 }, 00:08:25.140 "memory_domains": [ 00:08:25.140 { 00:08:25.140 "dma_device_id": "system", 00:08:25.140 "dma_device_type": 1 00:08:25.140 }, 00:08:25.140 { 00:08:25.140 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:25.140 "dma_device_type": 2 00:08:25.140 } 00:08:25.140 ], 00:08:25.140 "driver_specific": {} 00:08:25.140 } 00:08:25.140 ] 00:08:25.140 15:14:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:25.140 15:14:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:25.140 15:14:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:25.140 15:14:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:25.140 15:14:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 2 00:08:25.140 15:14:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:25.140 15:14:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:25.140 15:14:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:25.140 15:14:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:25.140 15:14:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:25.140 15:14:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:25.140 15:14:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:25.140 15:14:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:25.140 15:14:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:25.140 15:14:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:25.140 15:14:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:25.140 15:14:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:25.140 15:14:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.140 15:14:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:25.140 15:14:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:25.140 "name": "Existed_Raid", 00:08:25.140 "uuid": "6d7b0fe5-9c30-4c7d-b749-584a444e5c35", 00:08:25.140 "strip_size_kb": 64, 00:08:25.140 "state": "online", 00:08:25.140 "raid_level": "raid0", 00:08:25.140 "superblock": true, 00:08:25.140 "num_base_bdevs": 2, 00:08:25.140 "num_base_bdevs_discovered": 2, 00:08:25.140 "num_base_bdevs_operational": 2, 00:08:25.140 "base_bdevs_list": [ 00:08:25.140 { 00:08:25.140 "name": "BaseBdev1", 00:08:25.140 "uuid": "b2a14192-e58a-4d64-99c5-e8ecad417653", 00:08:25.140 "is_configured": true, 00:08:25.140 "data_offset": 2048, 00:08:25.140 "data_size": 63488 00:08:25.140 }, 00:08:25.140 { 00:08:25.140 "name": "BaseBdev2", 00:08:25.140 "uuid": "0932504a-7dc3-453d-973c-bc2504031e1e", 00:08:25.140 "is_configured": true, 00:08:25.140 "data_offset": 2048, 00:08:25.140 "data_size": 63488 00:08:25.140 } 00:08:25.140 ] 00:08:25.140 }' 00:08:25.140 15:14:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:25.140 15:14:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.709 15:14:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:08:25.709 15:14:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:25.709 15:14:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:25.709 15:14:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:25.709 15:14:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:08:25.709 15:14:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:25.709 15:14:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:25.709 15:14:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:25.709 15:14:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:25.709 15:14:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.709 [2024-11-19 15:14:15.791879] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:25.709 15:14:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:25.709 15:14:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:25.709 "name": "Existed_Raid", 00:08:25.709 "aliases": [ 00:08:25.709 "6d7b0fe5-9c30-4c7d-b749-584a444e5c35" 00:08:25.709 ], 00:08:25.709 "product_name": "Raid Volume", 00:08:25.709 "block_size": 512, 00:08:25.709 "num_blocks": 126976, 00:08:25.709 "uuid": "6d7b0fe5-9c30-4c7d-b749-584a444e5c35", 00:08:25.709 "assigned_rate_limits": { 00:08:25.709 "rw_ios_per_sec": 0, 00:08:25.709 "rw_mbytes_per_sec": 0, 00:08:25.709 "r_mbytes_per_sec": 0, 00:08:25.709 "w_mbytes_per_sec": 0 00:08:25.709 }, 00:08:25.709 "claimed": false, 00:08:25.709 "zoned": false, 00:08:25.709 "supported_io_types": { 00:08:25.709 "read": true, 00:08:25.709 "write": true, 00:08:25.709 "unmap": true, 00:08:25.709 "flush": true, 00:08:25.709 "reset": true, 00:08:25.709 "nvme_admin": false, 00:08:25.709 "nvme_io": false, 00:08:25.709 "nvme_io_md": false, 00:08:25.709 "write_zeroes": true, 00:08:25.709 "zcopy": false, 00:08:25.709 "get_zone_info": false, 00:08:25.709 "zone_management": false, 00:08:25.709 "zone_append": false, 00:08:25.709 "compare": false, 00:08:25.709 "compare_and_write": false, 00:08:25.709 "abort": false, 00:08:25.709 "seek_hole": false, 00:08:25.709 "seek_data": false, 00:08:25.709 "copy": false, 00:08:25.709 "nvme_iov_md": false 00:08:25.709 }, 00:08:25.709 "memory_domains": [ 00:08:25.709 { 00:08:25.709 "dma_device_id": "system", 00:08:25.709 "dma_device_type": 1 00:08:25.709 }, 00:08:25.709 { 00:08:25.710 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:25.710 "dma_device_type": 2 00:08:25.710 }, 00:08:25.710 { 00:08:25.710 "dma_device_id": "system", 00:08:25.710 "dma_device_type": 1 00:08:25.710 }, 00:08:25.710 { 00:08:25.710 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:25.710 "dma_device_type": 2 00:08:25.710 } 00:08:25.710 ], 00:08:25.710 "driver_specific": { 00:08:25.710 "raid": { 00:08:25.710 "uuid": "6d7b0fe5-9c30-4c7d-b749-584a444e5c35", 00:08:25.710 "strip_size_kb": 64, 00:08:25.710 "state": "online", 00:08:25.710 "raid_level": "raid0", 00:08:25.710 "superblock": true, 00:08:25.710 "num_base_bdevs": 2, 00:08:25.710 "num_base_bdevs_discovered": 2, 00:08:25.710 "num_base_bdevs_operational": 2, 00:08:25.710 "base_bdevs_list": [ 00:08:25.710 { 00:08:25.710 "name": "BaseBdev1", 00:08:25.710 "uuid": "b2a14192-e58a-4d64-99c5-e8ecad417653", 00:08:25.710 "is_configured": true, 00:08:25.710 "data_offset": 2048, 00:08:25.710 "data_size": 63488 00:08:25.710 }, 00:08:25.710 { 00:08:25.710 "name": "BaseBdev2", 00:08:25.710 "uuid": "0932504a-7dc3-453d-973c-bc2504031e1e", 00:08:25.710 "is_configured": true, 00:08:25.710 "data_offset": 2048, 00:08:25.710 "data_size": 63488 00:08:25.710 } 00:08:25.710 ] 00:08:25.710 } 00:08:25.710 } 00:08:25.710 }' 00:08:25.710 15:14:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:25.710 15:14:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:08:25.710 BaseBdev2' 00:08:25.710 15:14:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:25.710 15:14:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:25.710 15:14:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:25.710 15:14:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:08:25.710 15:14:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:25.710 15:14:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.710 15:14:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:25.710 15:14:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:25.710 15:14:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:25.710 15:14:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:25.710 15:14:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:25.710 15:14:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:25.710 15:14:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:25.710 15:14:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.710 15:14:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:25.710 15:14:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:25.710 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:25.710 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:25.710 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:25.710 15:14:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:25.710 15:14:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.710 [2024-11-19 15:14:16.011184] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:25.710 [2024-11-19 15:14:16.011311] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:25.710 [2024-11-19 15:14:16.011380] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:25.710 15:14:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:25.710 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:08:25.710 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:08:25.710 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:25.710 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:08:25.710 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:08:25.710 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 1 00:08:25.710 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:25.710 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:08:25.710 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:25.710 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:25.710 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:08:25.710 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:25.710 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:25.710 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:25.710 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:25.710 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:25.710 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:25.710 15:14:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:25.710 15:14:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.970 15:14:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:25.970 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:25.970 "name": "Existed_Raid", 00:08:25.970 "uuid": "6d7b0fe5-9c30-4c7d-b749-584a444e5c35", 00:08:25.970 "strip_size_kb": 64, 00:08:25.970 "state": "offline", 00:08:25.970 "raid_level": "raid0", 00:08:25.970 "superblock": true, 00:08:25.970 "num_base_bdevs": 2, 00:08:25.970 "num_base_bdevs_discovered": 1, 00:08:25.970 "num_base_bdevs_operational": 1, 00:08:25.970 "base_bdevs_list": [ 00:08:25.970 { 00:08:25.970 "name": null, 00:08:25.970 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:25.970 "is_configured": false, 00:08:25.970 "data_offset": 0, 00:08:25.970 "data_size": 63488 00:08:25.970 }, 00:08:25.970 { 00:08:25.970 "name": "BaseBdev2", 00:08:25.970 "uuid": "0932504a-7dc3-453d-973c-bc2504031e1e", 00:08:25.970 "is_configured": true, 00:08:25.970 "data_offset": 2048, 00:08:25.970 "data_size": 63488 00:08:25.970 } 00:08:25.970 ] 00:08:25.970 }' 00:08:25.970 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:25.970 15:14:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.230 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:08:26.230 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:26.230 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:26.230 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:26.230 15:14:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.230 15:14:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.230 15:14:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.230 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:26.230 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:26.230 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:08:26.230 15:14:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.230 15:14:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.230 [2024-11-19 15:14:16.503227] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:26.230 [2024-11-19 15:14:16.503298] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:08:26.230 15:14:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.230 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:26.230 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:26.230 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:08:26.230 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:26.230 15:14:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.230 15:14:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.230 15:14:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.490 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:08:26.490 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:08:26.490 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:08:26.490 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 72323 00:08:26.490 15:14:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 72323 ']' 00:08:26.490 15:14:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 72323 00:08:26.490 15:14:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:08:26.490 15:14:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:26.490 15:14:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 72323 00:08:26.490 killing process with pid 72323 00:08:26.490 15:14:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:26.490 15:14:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:26.490 15:14:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 72323' 00:08:26.490 15:14:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 72323 00:08:26.490 [2024-11-19 15:14:16.622806] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:26.490 15:14:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 72323 00:08:26.490 [2024-11-19 15:14:16.624429] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:26.750 15:14:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:08:26.750 00:08:26.750 real 0m4.042s 00:08:26.750 user 0m6.269s 00:08:26.750 sys 0m0.844s 00:08:26.750 ************************************ 00:08:26.750 END TEST raid_state_function_test_sb 00:08:26.750 ************************************ 00:08:26.750 15:14:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:26.750 15:14:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.750 15:14:17 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid0 2 00:08:26.750 15:14:17 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:08:26.750 15:14:17 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:26.750 15:14:17 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:26.750 ************************************ 00:08:26.750 START TEST raid_superblock_test 00:08:26.750 ************************************ 00:08:26.750 15:14:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid0 2 00:08:26.750 15:14:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid0 00:08:26.750 15:14:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:08:26.750 15:14:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:08:26.750 15:14:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:08:26.750 15:14:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:08:26.750 15:14:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:08:26.750 15:14:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:08:26.750 15:14:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:08:26.750 15:14:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:08:26.750 15:14:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:08:26.750 15:14:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:08:26.750 15:14:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:08:26.750 15:14:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:08:26.750 15:14:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid0 '!=' raid1 ']' 00:08:26.750 15:14:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:08:26.750 15:14:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:08:26.750 15:14:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=72564 00:08:26.750 15:14:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:08:26.750 15:14:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 72564 00:08:26.750 15:14:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 72564 ']' 00:08:26.750 15:14:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:26.750 15:14:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:26.750 15:14:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:26.750 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:26.750 15:14:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:26.750 15:14:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.010 [2024-11-19 15:14:17.101729] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:08:27.010 [2024-11-19 15:14:17.101873] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid72564 ] 00:08:27.010 [2024-11-19 15:14:17.242577] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:27.010 [2024-11-19 15:14:17.283335] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:27.270 [2024-11-19 15:14:17.361088] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:27.270 [2024-11-19 15:14:17.361128] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:27.840 15:14:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:27.840 15:14:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:08:27.840 15:14:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:08:27.840 15:14:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:27.840 15:14:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:08:27.840 15:14:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:08:27.840 15:14:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:08:27.840 15:14:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:27.840 15:14:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:27.840 15:14:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:27.840 15:14:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:08:27.840 15:14:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:27.840 15:14:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.840 malloc1 00:08:27.840 15:14:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:27.840 15:14:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:27.840 15:14:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:27.840 15:14:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.840 [2024-11-19 15:14:17.988054] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:27.840 [2024-11-19 15:14:17.988216] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:27.840 [2024-11-19 15:14:17.988259] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:08:27.840 [2024-11-19 15:14:17.988297] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:27.840 [2024-11-19 15:14:17.990823] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:27.840 [2024-11-19 15:14:17.990899] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:27.840 pt1 00:08:27.840 15:14:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:27.840 15:14:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:27.840 15:14:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:27.840 15:14:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:08:27.840 15:14:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:08:27.840 15:14:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:08:27.840 15:14:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:27.840 15:14:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:27.840 15:14:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:27.840 15:14:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:08:27.840 15:14:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:27.840 15:14:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.840 malloc2 00:08:27.840 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:27.840 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:27.840 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:27.840 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.840 [2024-11-19 15:14:18.026697] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:27.840 [2024-11-19 15:14:18.026753] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:27.840 [2024-11-19 15:14:18.026769] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:08:27.840 [2024-11-19 15:14:18.026781] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:27.840 [2024-11-19 15:14:18.029223] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:27.840 [2024-11-19 15:14:18.029258] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:27.840 pt2 00:08:27.840 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:27.840 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:27.840 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:27.840 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:08:27.840 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:27.840 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.840 [2024-11-19 15:14:18.038716] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:27.840 [2024-11-19 15:14:18.040896] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:27.840 [2024-11-19 15:14:18.041129] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:08:27.840 [2024-11-19 15:14:18.041149] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:08:27.841 [2024-11-19 15:14:18.041441] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:08:27.841 [2024-11-19 15:14:18.041591] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:08:27.841 [2024-11-19 15:14:18.041602] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:08:27.841 [2024-11-19 15:14:18.041729] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:27.841 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:27.841 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:08:27.841 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:27.841 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:27.841 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:27.841 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:27.841 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:27.841 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:27.841 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:27.841 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:27.841 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:27.841 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:27.841 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:27.841 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:27.841 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.841 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:27.841 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:27.841 "name": "raid_bdev1", 00:08:27.841 "uuid": "9f244161-cbc9-481c-8fa8-4cf11a165644", 00:08:27.841 "strip_size_kb": 64, 00:08:27.841 "state": "online", 00:08:27.841 "raid_level": "raid0", 00:08:27.841 "superblock": true, 00:08:27.841 "num_base_bdevs": 2, 00:08:27.841 "num_base_bdevs_discovered": 2, 00:08:27.841 "num_base_bdevs_operational": 2, 00:08:27.841 "base_bdevs_list": [ 00:08:27.841 { 00:08:27.841 "name": "pt1", 00:08:27.841 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:27.841 "is_configured": true, 00:08:27.841 "data_offset": 2048, 00:08:27.841 "data_size": 63488 00:08:27.841 }, 00:08:27.841 { 00:08:27.841 "name": "pt2", 00:08:27.841 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:27.841 "is_configured": true, 00:08:27.841 "data_offset": 2048, 00:08:27.841 "data_size": 63488 00:08:27.841 } 00:08:27.841 ] 00:08:27.841 }' 00:08:27.841 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:27.841 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.411 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:08:28.411 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:08:28.411 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:28.411 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:28.411 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:28.411 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:28.411 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:28.411 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:28.411 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.411 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.411 [2024-11-19 15:14:18.494353] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:28.411 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.411 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:28.411 "name": "raid_bdev1", 00:08:28.411 "aliases": [ 00:08:28.411 "9f244161-cbc9-481c-8fa8-4cf11a165644" 00:08:28.411 ], 00:08:28.411 "product_name": "Raid Volume", 00:08:28.411 "block_size": 512, 00:08:28.411 "num_blocks": 126976, 00:08:28.411 "uuid": "9f244161-cbc9-481c-8fa8-4cf11a165644", 00:08:28.411 "assigned_rate_limits": { 00:08:28.411 "rw_ios_per_sec": 0, 00:08:28.411 "rw_mbytes_per_sec": 0, 00:08:28.411 "r_mbytes_per_sec": 0, 00:08:28.411 "w_mbytes_per_sec": 0 00:08:28.411 }, 00:08:28.411 "claimed": false, 00:08:28.411 "zoned": false, 00:08:28.411 "supported_io_types": { 00:08:28.411 "read": true, 00:08:28.411 "write": true, 00:08:28.411 "unmap": true, 00:08:28.411 "flush": true, 00:08:28.411 "reset": true, 00:08:28.411 "nvme_admin": false, 00:08:28.411 "nvme_io": false, 00:08:28.411 "nvme_io_md": false, 00:08:28.411 "write_zeroes": true, 00:08:28.411 "zcopy": false, 00:08:28.411 "get_zone_info": false, 00:08:28.411 "zone_management": false, 00:08:28.411 "zone_append": false, 00:08:28.411 "compare": false, 00:08:28.411 "compare_and_write": false, 00:08:28.411 "abort": false, 00:08:28.411 "seek_hole": false, 00:08:28.411 "seek_data": false, 00:08:28.411 "copy": false, 00:08:28.411 "nvme_iov_md": false 00:08:28.411 }, 00:08:28.411 "memory_domains": [ 00:08:28.411 { 00:08:28.411 "dma_device_id": "system", 00:08:28.411 "dma_device_type": 1 00:08:28.411 }, 00:08:28.411 { 00:08:28.411 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:28.411 "dma_device_type": 2 00:08:28.411 }, 00:08:28.411 { 00:08:28.411 "dma_device_id": "system", 00:08:28.411 "dma_device_type": 1 00:08:28.411 }, 00:08:28.411 { 00:08:28.411 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:28.411 "dma_device_type": 2 00:08:28.411 } 00:08:28.411 ], 00:08:28.411 "driver_specific": { 00:08:28.411 "raid": { 00:08:28.411 "uuid": "9f244161-cbc9-481c-8fa8-4cf11a165644", 00:08:28.411 "strip_size_kb": 64, 00:08:28.411 "state": "online", 00:08:28.411 "raid_level": "raid0", 00:08:28.411 "superblock": true, 00:08:28.411 "num_base_bdevs": 2, 00:08:28.411 "num_base_bdevs_discovered": 2, 00:08:28.411 "num_base_bdevs_operational": 2, 00:08:28.411 "base_bdevs_list": [ 00:08:28.411 { 00:08:28.411 "name": "pt1", 00:08:28.411 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:28.411 "is_configured": true, 00:08:28.411 "data_offset": 2048, 00:08:28.411 "data_size": 63488 00:08:28.411 }, 00:08:28.411 { 00:08:28.411 "name": "pt2", 00:08:28.411 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:28.411 "is_configured": true, 00:08:28.411 "data_offset": 2048, 00:08:28.411 "data_size": 63488 00:08:28.411 } 00:08:28.411 ] 00:08:28.411 } 00:08:28.411 } 00:08:28.411 }' 00:08:28.411 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:28.411 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:08:28.411 pt2' 00:08:28.411 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:28.411 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:28.411 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:28.411 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:08:28.411 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:28.411 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.411 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.411 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.411 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:28.411 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:28.411 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:28.411 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:28.411 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:08:28.411 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.411 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.411 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.411 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:28.411 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:28.411 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:08:28.411 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:28.411 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.411 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.411 [2024-11-19 15:14:18.709859] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:28.411 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.671 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=9f244161-cbc9-481c-8fa8-4cf11a165644 00:08:28.671 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 9f244161-cbc9-481c-8fa8-4cf11a165644 ']' 00:08:28.671 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:28.671 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.671 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.671 [2024-11-19 15:14:18.757514] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:28.671 [2024-11-19 15:14:18.757650] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:28.671 [2024-11-19 15:14:18.757815] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:28.671 [2024-11-19 15:14:18.757914] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:28.671 [2024-11-19 15:14:18.757960] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:08:28.671 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.671 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:28.671 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:08:28.671 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.671 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.671 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.671 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:08:28.671 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:08:28.671 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:28.671 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:08:28.671 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.671 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.671 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.671 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:28.671 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:08:28.671 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.671 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.671 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.671 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:08:28.671 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:08:28.671 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.671 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.671 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.671 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:08:28.671 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:08:28.671 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:08:28.672 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:08:28.672 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:08:28.672 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:08:28.672 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:08:28.672 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:08:28.672 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:08:28.672 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.672 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.672 [2024-11-19 15:14:18.877263] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:08:28.672 [2024-11-19 15:14:18.879510] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:08:28.672 [2024-11-19 15:14:18.879617] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:08:28.672 [2024-11-19 15:14:18.879707] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:08:28.672 [2024-11-19 15:14:18.879797] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:28.672 [2024-11-19 15:14:18.879830] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:08:28.672 request: 00:08:28.672 { 00:08:28.672 "name": "raid_bdev1", 00:08:28.672 "raid_level": "raid0", 00:08:28.672 "base_bdevs": [ 00:08:28.672 "malloc1", 00:08:28.672 "malloc2" 00:08:28.672 ], 00:08:28.672 "strip_size_kb": 64, 00:08:28.672 "superblock": false, 00:08:28.672 "method": "bdev_raid_create", 00:08:28.672 "req_id": 1 00:08:28.672 } 00:08:28.672 Got JSON-RPC error response 00:08:28.672 response: 00:08:28.672 { 00:08:28.672 "code": -17, 00:08:28.672 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:08:28.672 } 00:08:28.672 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:08:28.672 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:08:28.672 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:08:28.672 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:08:28.672 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:08:28.672 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:28.672 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.672 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.672 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:08:28.672 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.672 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:08:28.672 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:08:28.672 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:28.672 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.672 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.672 [2024-11-19 15:14:18.945114] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:28.672 [2024-11-19 15:14:18.945207] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:28.672 [2024-11-19 15:14:18.945230] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:08:28.672 [2024-11-19 15:14:18.945239] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:28.672 [2024-11-19 15:14:18.947766] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:28.672 [2024-11-19 15:14:18.947800] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:28.672 [2024-11-19 15:14:18.947867] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:08:28.672 [2024-11-19 15:14:18.947906] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:28.672 pt1 00:08:28.672 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.672 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 2 00:08:28.672 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:28.672 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:28.672 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:28.672 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:28.672 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:28.672 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:28.672 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:28.672 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:28.672 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:28.672 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:28.672 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:28.672 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.672 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:28.672 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:28.672 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:28.672 "name": "raid_bdev1", 00:08:28.672 "uuid": "9f244161-cbc9-481c-8fa8-4cf11a165644", 00:08:28.672 "strip_size_kb": 64, 00:08:28.672 "state": "configuring", 00:08:28.672 "raid_level": "raid0", 00:08:28.672 "superblock": true, 00:08:28.672 "num_base_bdevs": 2, 00:08:28.672 "num_base_bdevs_discovered": 1, 00:08:28.672 "num_base_bdevs_operational": 2, 00:08:28.672 "base_bdevs_list": [ 00:08:28.672 { 00:08:28.672 "name": "pt1", 00:08:28.672 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:28.672 "is_configured": true, 00:08:28.672 "data_offset": 2048, 00:08:28.672 "data_size": 63488 00:08:28.672 }, 00:08:28.672 { 00:08:28.672 "name": null, 00:08:28.672 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:28.672 "is_configured": false, 00:08:28.672 "data_offset": 2048, 00:08:28.672 "data_size": 63488 00:08:28.672 } 00:08:28.672 ] 00:08:28.672 }' 00:08:28.672 15:14:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:28.672 15:14:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.242 15:14:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:08:29.242 15:14:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:08:29.242 15:14:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:29.242 15:14:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:29.242 15:14:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.242 15:14:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.242 [2024-11-19 15:14:19.428389] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:29.242 [2024-11-19 15:14:19.428592] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:29.242 [2024-11-19 15:14:19.428637] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:08:29.242 [2024-11-19 15:14:19.428688] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:29.242 [2024-11-19 15:14:19.429230] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:29.242 [2024-11-19 15:14:19.429300] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:29.242 [2024-11-19 15:14:19.429426] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:08:29.242 [2024-11-19 15:14:19.429484] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:29.242 [2024-11-19 15:14:19.429611] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:29.242 [2024-11-19 15:14:19.429647] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:08:29.242 [2024-11-19 15:14:19.429952] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:08:29.242 [2024-11-19 15:14:19.430130] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:29.242 [2024-11-19 15:14:19.430180] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:08:29.243 [2024-11-19 15:14:19.430327] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:29.243 pt2 00:08:29.243 15:14:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.243 15:14:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:08:29.243 15:14:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:29.243 15:14:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:08:29.243 15:14:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:29.243 15:14:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:29.243 15:14:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:29.243 15:14:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:29.243 15:14:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:29.243 15:14:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:29.243 15:14:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:29.243 15:14:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:29.243 15:14:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:29.243 15:14:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:29.243 15:14:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.243 15:14:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.243 15:14:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:29.243 15:14:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.243 15:14:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:29.243 "name": "raid_bdev1", 00:08:29.243 "uuid": "9f244161-cbc9-481c-8fa8-4cf11a165644", 00:08:29.243 "strip_size_kb": 64, 00:08:29.243 "state": "online", 00:08:29.243 "raid_level": "raid0", 00:08:29.243 "superblock": true, 00:08:29.243 "num_base_bdevs": 2, 00:08:29.243 "num_base_bdevs_discovered": 2, 00:08:29.243 "num_base_bdevs_operational": 2, 00:08:29.243 "base_bdevs_list": [ 00:08:29.243 { 00:08:29.243 "name": "pt1", 00:08:29.243 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:29.243 "is_configured": true, 00:08:29.243 "data_offset": 2048, 00:08:29.243 "data_size": 63488 00:08:29.243 }, 00:08:29.243 { 00:08:29.243 "name": "pt2", 00:08:29.243 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:29.243 "is_configured": true, 00:08:29.243 "data_offset": 2048, 00:08:29.243 "data_size": 63488 00:08:29.243 } 00:08:29.243 ] 00:08:29.243 }' 00:08:29.243 15:14:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:29.243 15:14:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.821 15:14:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:08:29.821 15:14:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:08:29.821 15:14:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:29.821 15:14:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:29.821 15:14:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:29.821 15:14:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:29.821 15:14:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:29.821 15:14:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:29.821 15:14:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.821 15:14:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.821 [2024-11-19 15:14:19.875930] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:29.821 15:14:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.821 15:14:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:29.821 "name": "raid_bdev1", 00:08:29.821 "aliases": [ 00:08:29.821 "9f244161-cbc9-481c-8fa8-4cf11a165644" 00:08:29.821 ], 00:08:29.821 "product_name": "Raid Volume", 00:08:29.821 "block_size": 512, 00:08:29.821 "num_blocks": 126976, 00:08:29.821 "uuid": "9f244161-cbc9-481c-8fa8-4cf11a165644", 00:08:29.821 "assigned_rate_limits": { 00:08:29.821 "rw_ios_per_sec": 0, 00:08:29.821 "rw_mbytes_per_sec": 0, 00:08:29.821 "r_mbytes_per_sec": 0, 00:08:29.821 "w_mbytes_per_sec": 0 00:08:29.821 }, 00:08:29.821 "claimed": false, 00:08:29.821 "zoned": false, 00:08:29.821 "supported_io_types": { 00:08:29.821 "read": true, 00:08:29.821 "write": true, 00:08:29.821 "unmap": true, 00:08:29.821 "flush": true, 00:08:29.821 "reset": true, 00:08:29.821 "nvme_admin": false, 00:08:29.821 "nvme_io": false, 00:08:29.821 "nvme_io_md": false, 00:08:29.821 "write_zeroes": true, 00:08:29.821 "zcopy": false, 00:08:29.821 "get_zone_info": false, 00:08:29.821 "zone_management": false, 00:08:29.821 "zone_append": false, 00:08:29.821 "compare": false, 00:08:29.821 "compare_and_write": false, 00:08:29.821 "abort": false, 00:08:29.821 "seek_hole": false, 00:08:29.821 "seek_data": false, 00:08:29.821 "copy": false, 00:08:29.821 "nvme_iov_md": false 00:08:29.821 }, 00:08:29.821 "memory_domains": [ 00:08:29.821 { 00:08:29.821 "dma_device_id": "system", 00:08:29.821 "dma_device_type": 1 00:08:29.821 }, 00:08:29.821 { 00:08:29.821 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:29.821 "dma_device_type": 2 00:08:29.821 }, 00:08:29.821 { 00:08:29.821 "dma_device_id": "system", 00:08:29.821 "dma_device_type": 1 00:08:29.821 }, 00:08:29.821 { 00:08:29.821 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:29.821 "dma_device_type": 2 00:08:29.821 } 00:08:29.821 ], 00:08:29.821 "driver_specific": { 00:08:29.821 "raid": { 00:08:29.821 "uuid": "9f244161-cbc9-481c-8fa8-4cf11a165644", 00:08:29.821 "strip_size_kb": 64, 00:08:29.821 "state": "online", 00:08:29.821 "raid_level": "raid0", 00:08:29.821 "superblock": true, 00:08:29.821 "num_base_bdevs": 2, 00:08:29.821 "num_base_bdevs_discovered": 2, 00:08:29.821 "num_base_bdevs_operational": 2, 00:08:29.821 "base_bdevs_list": [ 00:08:29.821 { 00:08:29.821 "name": "pt1", 00:08:29.821 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:29.821 "is_configured": true, 00:08:29.821 "data_offset": 2048, 00:08:29.821 "data_size": 63488 00:08:29.821 }, 00:08:29.821 { 00:08:29.821 "name": "pt2", 00:08:29.821 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:29.821 "is_configured": true, 00:08:29.821 "data_offset": 2048, 00:08:29.821 "data_size": 63488 00:08:29.821 } 00:08:29.821 ] 00:08:29.821 } 00:08:29.821 } 00:08:29.821 }' 00:08:29.821 15:14:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:29.821 15:14:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:08:29.821 pt2' 00:08:29.821 15:14:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:29.821 15:14:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:29.821 15:14:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:29.821 15:14:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:08:29.821 15:14:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.821 15:14:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.821 15:14:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:29.821 15:14:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.821 15:14:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:29.821 15:14:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:29.821 15:14:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:29.821 15:14:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:08:29.821 15:14:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.821 15:14:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.821 15:14:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:29.821 15:14:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.821 15:14:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:29.821 15:14:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:29.822 15:14:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:29.822 15:14:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.822 15:14:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.822 15:14:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:08:29.822 [2024-11-19 15:14:20.087583] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:29.822 15:14:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.822 15:14:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 9f244161-cbc9-481c-8fa8-4cf11a165644 '!=' 9f244161-cbc9-481c-8fa8-4cf11a165644 ']' 00:08:29.822 15:14:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid0 00:08:29.822 15:14:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:29.822 15:14:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:29.822 15:14:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 72564 00:08:29.822 15:14:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 72564 ']' 00:08:29.822 15:14:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 72564 00:08:29.822 15:14:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:08:29.822 15:14:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:29.822 15:14:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 72564 00:08:30.082 15:14:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:30.082 15:14:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:30.082 15:14:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 72564' 00:08:30.082 killing process with pid 72564 00:08:30.082 15:14:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 72564 00:08:30.082 [2024-11-19 15:14:20.177411] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:30.082 15:14:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 72564 00:08:30.082 [2024-11-19 15:14:20.177607] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:30.082 [2024-11-19 15:14:20.177667] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:30.082 [2024-11-19 15:14:20.177677] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:08:30.082 [2024-11-19 15:14:20.218607] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:30.342 15:14:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:08:30.342 00:08:30.342 real 0m3.522s 00:08:30.342 user 0m5.287s 00:08:30.342 sys 0m0.802s 00:08:30.342 15:14:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:30.342 15:14:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:30.342 ************************************ 00:08:30.342 END TEST raid_superblock_test 00:08:30.342 ************************************ 00:08:30.342 15:14:20 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid0 2 read 00:08:30.342 15:14:20 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:30.342 15:14:20 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:30.342 15:14:20 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:30.342 ************************************ 00:08:30.342 START TEST raid_read_error_test 00:08:30.342 ************************************ 00:08:30.342 15:14:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid0 2 read 00:08:30.342 15:14:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:08:30.342 15:14:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:08:30.342 15:14:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:08:30.342 15:14:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:08:30.342 15:14:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:30.342 15:14:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:08:30.342 15:14:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:30.342 15:14:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:30.342 15:14:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:08:30.342 15:14:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:30.342 15:14:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:30.342 15:14:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:08:30.342 15:14:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:08:30.342 15:14:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:08:30.342 15:14:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:08:30.342 15:14:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:08:30.342 15:14:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:08:30.342 15:14:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:08:30.342 15:14:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:08:30.342 15:14:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:08:30.342 15:14:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:08:30.342 15:14:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:08:30.342 15:14:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.SfEjODweCP 00:08:30.342 15:14:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=72770 00:08:30.342 15:14:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:08:30.342 15:14:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 72770 00:08:30.342 15:14:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 72770 ']' 00:08:30.342 15:14:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:30.342 15:14:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:30.342 15:14:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:30.342 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:30.342 15:14:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:30.342 15:14:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:30.602 [2024-11-19 15:14:20.711723] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:08:30.602 [2024-11-19 15:14:20.711858] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid72770 ] 00:08:30.602 [2024-11-19 15:14:20.867913] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:30.602 [2024-11-19 15:14:20.909294] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:30.861 [2024-11-19 15:14:20.986256] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:30.861 [2024-11-19 15:14:20.986304] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.432 BaseBdev1_malloc 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.432 true 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.432 [2024-11-19 15:14:21.576780] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:08:31.432 [2024-11-19 15:14:21.576855] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:31.432 [2024-11-19 15:14:21.576881] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:08:31.432 [2024-11-19 15:14:21.576899] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:31.432 [2024-11-19 15:14:21.579433] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:31.432 [2024-11-19 15:14:21.579562] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:08:31.432 BaseBdev1 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.432 BaseBdev2_malloc 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.432 true 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.432 [2024-11-19 15:14:21.623528] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:08:31.432 [2024-11-19 15:14:21.623587] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:31.432 [2024-11-19 15:14:21.623608] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:08:31.432 [2024-11-19 15:14:21.623627] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:31.432 [2024-11-19 15:14:21.626075] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:31.432 [2024-11-19 15:14:21.626113] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:08:31.432 BaseBdev2 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.432 [2024-11-19 15:14:21.635587] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:31.432 [2024-11-19 15:14:21.637772] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:31.432 [2024-11-19 15:14:21.637995] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:31.432 [2024-11-19 15:14:21.638010] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:08:31.432 [2024-11-19 15:14:21.638290] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:08:31.432 [2024-11-19 15:14:21.638437] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:31.432 [2024-11-19 15:14:21.638449] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:08:31.432 [2024-11-19 15:14:21.638599] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:31.432 "name": "raid_bdev1", 00:08:31.432 "uuid": "992093ce-8c00-4597-ba5c-693aed8e7990", 00:08:31.432 "strip_size_kb": 64, 00:08:31.432 "state": "online", 00:08:31.432 "raid_level": "raid0", 00:08:31.432 "superblock": true, 00:08:31.432 "num_base_bdevs": 2, 00:08:31.432 "num_base_bdevs_discovered": 2, 00:08:31.432 "num_base_bdevs_operational": 2, 00:08:31.432 "base_bdevs_list": [ 00:08:31.432 { 00:08:31.432 "name": "BaseBdev1", 00:08:31.432 "uuid": "074ee617-c2aa-558b-b6be-c22091e0e2ed", 00:08:31.432 "is_configured": true, 00:08:31.432 "data_offset": 2048, 00:08:31.432 "data_size": 63488 00:08:31.432 }, 00:08:31.432 { 00:08:31.432 "name": "BaseBdev2", 00:08:31.432 "uuid": "0ab40aa0-cd40-5cdb-8213-ad7d78b9cfc9", 00:08:31.432 "is_configured": true, 00:08:31.432 "data_offset": 2048, 00:08:31.432 "data_size": 63488 00:08:31.432 } 00:08:31.432 ] 00:08:31.432 }' 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:31.432 15:14:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:32.003 15:14:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:08:32.003 15:14:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:08:32.003 [2024-11-19 15:14:22.115293] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:08:32.944 15:14:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:08:32.944 15:14:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:32.944 15:14:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:32.944 15:14:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:32.944 15:14:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:08:32.944 15:14:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:08:32.944 15:14:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:08:32.944 15:14:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:08:32.944 15:14:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:32.944 15:14:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:32.944 15:14:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:32.944 15:14:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:32.944 15:14:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:32.945 15:14:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:32.945 15:14:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:32.945 15:14:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:32.945 15:14:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:32.945 15:14:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:32.945 15:14:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:32.945 15:14:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:32.945 15:14:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:32.945 15:14:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:32.945 15:14:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:32.945 "name": "raid_bdev1", 00:08:32.945 "uuid": "992093ce-8c00-4597-ba5c-693aed8e7990", 00:08:32.945 "strip_size_kb": 64, 00:08:32.945 "state": "online", 00:08:32.945 "raid_level": "raid0", 00:08:32.945 "superblock": true, 00:08:32.945 "num_base_bdevs": 2, 00:08:32.945 "num_base_bdevs_discovered": 2, 00:08:32.945 "num_base_bdevs_operational": 2, 00:08:32.945 "base_bdevs_list": [ 00:08:32.945 { 00:08:32.945 "name": "BaseBdev1", 00:08:32.945 "uuid": "074ee617-c2aa-558b-b6be-c22091e0e2ed", 00:08:32.945 "is_configured": true, 00:08:32.945 "data_offset": 2048, 00:08:32.945 "data_size": 63488 00:08:32.945 }, 00:08:32.945 { 00:08:32.945 "name": "BaseBdev2", 00:08:32.945 "uuid": "0ab40aa0-cd40-5cdb-8213-ad7d78b9cfc9", 00:08:32.945 "is_configured": true, 00:08:32.945 "data_offset": 2048, 00:08:32.945 "data_size": 63488 00:08:32.945 } 00:08:32.945 ] 00:08:32.945 }' 00:08:32.945 15:14:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:32.945 15:14:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.205 15:14:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:33.205 15:14:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:33.205 15:14:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.205 [2024-11-19 15:14:23.504184] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:33.205 [2024-11-19 15:14:23.504337] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:33.205 [2024-11-19 15:14:23.506853] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:33.205 [2024-11-19 15:14:23.506991] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:33.205 [2024-11-19 15:14:23.507055] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:33.206 [2024-11-19 15:14:23.507121] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:08:33.206 { 00:08:33.206 "results": [ 00:08:33.206 { 00:08:33.206 "job": "raid_bdev1", 00:08:33.206 "core_mask": "0x1", 00:08:33.206 "workload": "randrw", 00:08:33.206 "percentage": 50, 00:08:33.206 "status": "finished", 00:08:33.206 "queue_depth": 1, 00:08:33.206 "io_size": 131072, 00:08:33.206 "runtime": 1.389454, 00:08:33.206 "iops": 14728.087435784128, 00:08:33.206 "mibps": 1841.010929473016, 00:08:33.206 "io_failed": 1, 00:08:33.206 "io_timeout": 0, 00:08:33.206 "avg_latency_us": 95.27607048779629, 00:08:33.206 "min_latency_us": 24.817467248908297, 00:08:33.206 "max_latency_us": 1430.9170305676855 00:08:33.206 } 00:08:33.206 ], 00:08:33.206 "core_count": 1 00:08:33.206 } 00:08:33.206 15:14:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:33.206 15:14:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 72770 00:08:33.206 15:14:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 72770 ']' 00:08:33.206 15:14:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 72770 00:08:33.206 15:14:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:08:33.206 15:14:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:33.206 15:14:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 72770 00:08:33.466 killing process with pid 72770 00:08:33.466 15:14:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:33.466 15:14:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:33.466 15:14:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 72770' 00:08:33.466 15:14:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 72770 00:08:33.466 [2024-11-19 15:14:23.555337] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:33.466 15:14:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 72770 00:08:33.466 [2024-11-19 15:14:23.585801] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:33.727 15:14:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.SfEjODweCP 00:08:33.727 15:14:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:08:33.727 15:14:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:08:33.727 15:14:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.72 00:08:33.727 15:14:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:08:33.727 15:14:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:33.727 15:14:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:33.727 ************************************ 00:08:33.727 END TEST raid_read_error_test 00:08:33.727 ************************************ 00:08:33.727 15:14:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.72 != \0\.\0\0 ]] 00:08:33.727 00:08:33.727 real 0m3.305s 00:08:33.727 user 0m4.083s 00:08:33.727 sys 0m0.563s 00:08:33.727 15:14:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:33.727 15:14:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.727 15:14:23 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid0 2 write 00:08:33.727 15:14:23 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:33.727 15:14:23 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:33.727 15:14:23 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:33.727 ************************************ 00:08:33.727 START TEST raid_write_error_test 00:08:33.727 ************************************ 00:08:33.727 15:14:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid0 2 write 00:08:33.727 15:14:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:08:33.727 15:14:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:08:33.727 15:14:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:08:33.727 15:14:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:08:33.727 15:14:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:33.727 15:14:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:08:33.727 15:14:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:33.727 15:14:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:33.727 15:14:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:08:33.727 15:14:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:33.727 15:14:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:33.727 15:14:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:08:33.727 15:14:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:08:33.727 15:14:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:08:33.727 15:14:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:08:33.727 15:14:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:08:33.727 15:14:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:08:33.727 15:14:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:08:33.727 15:14:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:08:33.727 15:14:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:08:33.727 15:14:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:08:33.727 15:14:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:08:33.727 15:14:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.JF7Dk9F3B4 00:08:33.727 15:14:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=72899 00:08:33.727 15:14:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:08:33.727 15:14:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 72899 00:08:33.727 15:14:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 72899 ']' 00:08:33.727 15:14:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:33.727 15:14:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:33.727 15:14:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:33.727 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:33.727 15:14:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:33.727 15:14:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.987 [2024-11-19 15:14:24.080471] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:08:33.987 [2024-11-19 15:14:24.080604] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid72899 ] 00:08:33.987 [2024-11-19 15:14:24.216176] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:33.987 [2024-11-19 15:14:24.257113] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:34.247 [2024-11-19 15:14:24.334734] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:34.247 [2024-11-19 15:14:24.334910] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:34.818 15:14:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:34.818 15:14:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:08:34.818 15:14:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:34.818 15:14:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:08:34.818 15:14:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:34.818 15:14:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.818 BaseBdev1_malloc 00:08:34.818 15:14:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:34.818 15:14:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:08:34.818 15:14:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:34.818 15:14:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.818 true 00:08:34.818 15:14:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:34.818 15:14:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:08:34.818 15:14:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:34.818 15:14:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.818 [2024-11-19 15:14:24.990429] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:08:34.818 [2024-11-19 15:14:24.990610] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:34.818 [2024-11-19 15:14:24.990658] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:08:34.818 [2024-11-19 15:14:24.990698] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:34.818 [2024-11-19 15:14:24.993207] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:34.818 [2024-11-19 15:14:24.993281] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:08:34.818 BaseBdev1 00:08:34.818 15:14:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:34.818 15:14:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:34.818 15:14:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:08:34.818 15:14:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:34.818 15:14:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.818 BaseBdev2_malloc 00:08:34.818 15:14:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:34.818 15:14:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:08:34.818 15:14:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:34.818 15:14:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.818 true 00:08:34.818 15:14:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:34.818 15:14:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:08:34.818 15:14:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:34.818 15:14:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.818 [2024-11-19 15:14:25.037112] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:08:34.818 [2024-11-19 15:14:25.037173] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:34.818 [2024-11-19 15:14:25.037193] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:08:34.818 [2024-11-19 15:14:25.037212] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:34.818 [2024-11-19 15:14:25.039629] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:34.818 [2024-11-19 15:14:25.039748] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:08:34.818 BaseBdev2 00:08:34.818 15:14:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:34.818 15:14:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:08:34.818 15:14:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:34.818 15:14:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.818 [2024-11-19 15:14:25.049149] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:34.818 [2024-11-19 15:14:25.051368] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:34.818 [2024-11-19 15:14:25.051557] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:34.818 [2024-11-19 15:14:25.051571] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:08:34.818 [2024-11-19 15:14:25.051902] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:08:34.818 [2024-11-19 15:14:25.052094] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:34.818 [2024-11-19 15:14:25.052110] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:08:34.818 [2024-11-19 15:14:25.052243] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:34.818 15:14:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:34.818 15:14:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:08:34.818 15:14:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:34.818 15:14:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:34.818 15:14:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:34.818 15:14:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:34.818 15:14:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:34.818 15:14:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:34.818 15:14:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:34.819 15:14:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:34.819 15:14:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:34.819 15:14:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:34.819 15:14:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:34.819 15:14:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:34.819 15:14:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.819 15:14:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:34.819 15:14:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:34.819 "name": "raid_bdev1", 00:08:34.819 "uuid": "1096b1de-0b03-4156-8f80-1e0d17ac6bd0", 00:08:34.819 "strip_size_kb": 64, 00:08:34.819 "state": "online", 00:08:34.819 "raid_level": "raid0", 00:08:34.819 "superblock": true, 00:08:34.819 "num_base_bdevs": 2, 00:08:34.819 "num_base_bdevs_discovered": 2, 00:08:34.819 "num_base_bdevs_operational": 2, 00:08:34.819 "base_bdevs_list": [ 00:08:34.819 { 00:08:34.819 "name": "BaseBdev1", 00:08:34.819 "uuid": "69052350-ca15-53f9-9bcd-8ee8057be956", 00:08:34.819 "is_configured": true, 00:08:34.819 "data_offset": 2048, 00:08:34.819 "data_size": 63488 00:08:34.819 }, 00:08:34.819 { 00:08:34.819 "name": "BaseBdev2", 00:08:34.819 "uuid": "ec955b1f-9626-5432-858f-33fe51257b3c", 00:08:34.819 "is_configured": true, 00:08:34.819 "data_offset": 2048, 00:08:34.819 "data_size": 63488 00:08:34.819 } 00:08:34.819 ] 00:08:34.819 }' 00:08:34.819 15:14:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:34.819 15:14:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:35.389 15:14:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:08:35.389 15:14:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:08:35.389 [2024-11-19 15:14:25.576772] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:08:36.330 15:14:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:08:36.330 15:14:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:36.330 15:14:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.330 15:14:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:36.330 15:14:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:08:36.330 15:14:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:08:36.330 15:14:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:08:36.330 15:14:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:08:36.330 15:14:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:36.330 15:14:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:36.330 15:14:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:36.330 15:14:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:36.330 15:14:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:36.330 15:14:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:36.330 15:14:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:36.330 15:14:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:36.330 15:14:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:36.330 15:14:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:36.330 15:14:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:36.330 15:14:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:36.330 15:14:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.330 15:14:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:36.330 15:14:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:36.330 "name": "raid_bdev1", 00:08:36.330 "uuid": "1096b1de-0b03-4156-8f80-1e0d17ac6bd0", 00:08:36.330 "strip_size_kb": 64, 00:08:36.330 "state": "online", 00:08:36.330 "raid_level": "raid0", 00:08:36.330 "superblock": true, 00:08:36.330 "num_base_bdevs": 2, 00:08:36.330 "num_base_bdevs_discovered": 2, 00:08:36.330 "num_base_bdevs_operational": 2, 00:08:36.330 "base_bdevs_list": [ 00:08:36.330 { 00:08:36.330 "name": "BaseBdev1", 00:08:36.330 "uuid": "69052350-ca15-53f9-9bcd-8ee8057be956", 00:08:36.330 "is_configured": true, 00:08:36.330 "data_offset": 2048, 00:08:36.330 "data_size": 63488 00:08:36.330 }, 00:08:36.330 { 00:08:36.330 "name": "BaseBdev2", 00:08:36.330 "uuid": "ec955b1f-9626-5432-858f-33fe51257b3c", 00:08:36.330 "is_configured": true, 00:08:36.330 "data_offset": 2048, 00:08:36.330 "data_size": 63488 00:08:36.330 } 00:08:36.330 ] 00:08:36.330 }' 00:08:36.330 15:14:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:36.330 15:14:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.590 15:14:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:36.590 15:14:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:36.590 15:14:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.590 [2024-11-19 15:14:26.876833] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:36.590 [2024-11-19 15:14:26.876887] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:36.590 [2024-11-19 15:14:26.879421] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:36.590 { 00:08:36.590 "results": [ 00:08:36.590 { 00:08:36.590 "job": "raid_bdev1", 00:08:36.590 "core_mask": "0x1", 00:08:36.590 "workload": "randrw", 00:08:36.590 "percentage": 50, 00:08:36.590 "status": "finished", 00:08:36.590 "queue_depth": 1, 00:08:36.590 "io_size": 131072, 00:08:36.590 "runtime": 1.300306, 00:08:36.590 "iops": 14653.473874611052, 00:08:36.590 "mibps": 1831.6842343263816, 00:08:36.590 "io_failed": 1, 00:08:36.590 "io_timeout": 0, 00:08:36.590 "avg_latency_us": 95.69866607693886, 00:08:36.590 "min_latency_us": 25.7117903930131, 00:08:36.590 "max_latency_us": 1352.216593886463 00:08:36.590 } 00:08:36.590 ], 00:08:36.590 "core_count": 1 00:08:36.590 } 00:08:36.590 [2024-11-19 15:14:26.879593] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:36.590 [2024-11-19 15:14:26.879643] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:36.590 [2024-11-19 15:14:26.879653] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:08:36.590 15:14:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:36.590 15:14:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 72899 00:08:36.590 15:14:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 72899 ']' 00:08:36.590 15:14:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 72899 00:08:36.590 15:14:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:08:36.590 15:14:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:36.590 15:14:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 72899 00:08:36.590 killing process with pid 72899 00:08:36.590 15:14:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:36.590 15:14:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:36.590 15:14:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 72899' 00:08:36.590 15:14:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 72899 00:08:36.590 [2024-11-19 15:14:26.925333] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:36.590 15:14:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 72899 00:08:36.850 [2024-11-19 15:14:26.954832] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:37.110 15:14:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.JF7Dk9F3B4 00:08:37.110 15:14:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:08:37.110 15:14:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:08:37.110 15:14:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.77 00:08:37.110 15:14:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:08:37.110 15:14:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:37.110 15:14:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:37.110 15:14:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.77 != \0\.\0\0 ]] 00:08:37.110 00:08:37.110 real 0m3.301s 00:08:37.110 user 0m4.067s 00:08:37.110 sys 0m0.581s 00:08:37.110 15:14:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:37.110 15:14:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:37.110 ************************************ 00:08:37.110 END TEST raid_write_error_test 00:08:37.110 ************************************ 00:08:37.110 15:14:27 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:08:37.110 15:14:27 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test concat 2 false 00:08:37.110 15:14:27 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:37.110 15:14:27 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:37.110 15:14:27 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:37.110 ************************************ 00:08:37.110 START TEST raid_state_function_test 00:08:37.110 ************************************ 00:08:37.110 15:14:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test concat 2 false 00:08:37.110 15:14:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:08:37.110 15:14:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:08:37.110 15:14:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:08:37.110 15:14:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:37.110 15:14:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:37.110 15:14:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:37.110 15:14:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:37.110 15:14:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:37.110 15:14:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:37.110 15:14:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:37.110 15:14:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:37.110 15:14:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:37.110 15:14:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:08:37.110 15:14:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:37.110 15:14:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:37.110 15:14:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:37.110 15:14:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:37.110 15:14:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:37.110 15:14:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:08:37.110 15:14:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:08:37.110 15:14:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:08:37.110 15:14:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:08:37.110 15:14:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:08:37.110 15:14:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=73026 00:08:37.110 15:14:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:37.110 15:14:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 73026' 00:08:37.110 Process raid pid: 73026 00:08:37.110 15:14:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 73026 00:08:37.110 15:14:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 73026 ']' 00:08:37.110 15:14:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:37.110 15:14:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:37.110 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:37.110 15:14:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:37.110 15:14:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:37.110 15:14:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:37.110 [2024-11-19 15:14:27.443143] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:08:37.110 [2024-11-19 15:14:27.443284] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:37.369 [2024-11-19 15:14:27.601983] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:37.369 [2024-11-19 15:14:27.645141] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:37.629 [2024-11-19 15:14:27.722812] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:37.629 [2024-11-19 15:14:27.722863] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:38.197 15:14:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:38.197 15:14:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:08:38.197 15:14:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:08:38.197 15:14:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:38.197 15:14:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:38.197 [2024-11-19 15:14:28.282586] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:38.197 [2024-11-19 15:14:28.282665] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:38.197 [2024-11-19 15:14:28.282675] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:38.197 [2024-11-19 15:14:28.282685] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:38.197 15:14:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:38.197 15:14:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:08:38.197 15:14:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:38.197 15:14:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:38.197 15:14:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:38.197 15:14:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:38.197 15:14:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:38.197 15:14:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:38.197 15:14:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:38.197 15:14:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:38.197 15:14:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:38.197 15:14:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:38.197 15:14:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:38.197 15:14:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:38.197 15:14:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:38.197 15:14:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:38.197 15:14:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:38.197 "name": "Existed_Raid", 00:08:38.197 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:38.197 "strip_size_kb": 64, 00:08:38.197 "state": "configuring", 00:08:38.197 "raid_level": "concat", 00:08:38.197 "superblock": false, 00:08:38.197 "num_base_bdevs": 2, 00:08:38.197 "num_base_bdevs_discovered": 0, 00:08:38.197 "num_base_bdevs_operational": 2, 00:08:38.197 "base_bdevs_list": [ 00:08:38.197 { 00:08:38.197 "name": "BaseBdev1", 00:08:38.197 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:38.197 "is_configured": false, 00:08:38.197 "data_offset": 0, 00:08:38.197 "data_size": 0 00:08:38.197 }, 00:08:38.197 { 00:08:38.197 "name": "BaseBdev2", 00:08:38.197 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:38.197 "is_configured": false, 00:08:38.197 "data_offset": 0, 00:08:38.197 "data_size": 0 00:08:38.197 } 00:08:38.197 ] 00:08:38.197 }' 00:08:38.197 15:14:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:38.197 15:14:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:38.457 15:14:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:38.457 15:14:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:38.457 15:14:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:38.457 [2024-11-19 15:14:28.749768] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:38.457 [2024-11-19 15:14:28.749844] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:08:38.457 15:14:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:38.458 15:14:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:08:38.458 15:14:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:38.458 15:14:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:38.458 [2024-11-19 15:14:28.761732] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:38.458 [2024-11-19 15:14:28.761782] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:38.458 [2024-11-19 15:14:28.761791] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:38.458 [2024-11-19 15:14:28.761814] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:38.458 15:14:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:38.458 15:14:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:38.458 15:14:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:38.458 15:14:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:38.458 [2024-11-19 15:14:28.789164] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:38.458 BaseBdev1 00:08:38.458 15:14:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:38.458 15:14:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:08:38.458 15:14:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:08:38.458 15:14:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:38.458 15:14:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:38.458 15:14:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:38.458 15:14:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:38.458 15:14:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:38.458 15:14:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:38.458 15:14:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:38.719 15:14:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:38.719 15:14:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:38.719 15:14:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:38.719 15:14:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:38.719 [ 00:08:38.719 { 00:08:38.719 "name": "BaseBdev1", 00:08:38.719 "aliases": [ 00:08:38.719 "246a806b-c2c2-49ef-9243-13d444889ff1" 00:08:38.719 ], 00:08:38.719 "product_name": "Malloc disk", 00:08:38.719 "block_size": 512, 00:08:38.719 "num_blocks": 65536, 00:08:38.719 "uuid": "246a806b-c2c2-49ef-9243-13d444889ff1", 00:08:38.719 "assigned_rate_limits": { 00:08:38.719 "rw_ios_per_sec": 0, 00:08:38.719 "rw_mbytes_per_sec": 0, 00:08:38.719 "r_mbytes_per_sec": 0, 00:08:38.719 "w_mbytes_per_sec": 0 00:08:38.719 }, 00:08:38.719 "claimed": true, 00:08:38.719 "claim_type": "exclusive_write", 00:08:38.719 "zoned": false, 00:08:38.719 "supported_io_types": { 00:08:38.719 "read": true, 00:08:38.719 "write": true, 00:08:38.719 "unmap": true, 00:08:38.719 "flush": true, 00:08:38.719 "reset": true, 00:08:38.719 "nvme_admin": false, 00:08:38.719 "nvme_io": false, 00:08:38.719 "nvme_io_md": false, 00:08:38.719 "write_zeroes": true, 00:08:38.719 "zcopy": true, 00:08:38.719 "get_zone_info": false, 00:08:38.719 "zone_management": false, 00:08:38.719 "zone_append": false, 00:08:38.719 "compare": false, 00:08:38.719 "compare_and_write": false, 00:08:38.719 "abort": true, 00:08:38.719 "seek_hole": false, 00:08:38.719 "seek_data": false, 00:08:38.719 "copy": true, 00:08:38.719 "nvme_iov_md": false 00:08:38.719 }, 00:08:38.719 "memory_domains": [ 00:08:38.719 { 00:08:38.719 "dma_device_id": "system", 00:08:38.719 "dma_device_type": 1 00:08:38.719 }, 00:08:38.719 { 00:08:38.719 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:38.719 "dma_device_type": 2 00:08:38.719 } 00:08:38.719 ], 00:08:38.719 "driver_specific": {} 00:08:38.719 } 00:08:38.719 ] 00:08:38.719 15:14:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:38.719 15:14:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:38.719 15:14:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:08:38.719 15:14:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:38.719 15:14:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:38.719 15:14:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:38.719 15:14:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:38.719 15:14:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:38.719 15:14:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:38.719 15:14:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:38.719 15:14:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:38.719 15:14:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:38.719 15:14:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:38.719 15:14:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:38.719 15:14:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:38.719 15:14:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:38.719 15:14:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:38.719 15:14:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:38.719 "name": "Existed_Raid", 00:08:38.719 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:38.719 "strip_size_kb": 64, 00:08:38.719 "state": "configuring", 00:08:38.719 "raid_level": "concat", 00:08:38.719 "superblock": false, 00:08:38.719 "num_base_bdevs": 2, 00:08:38.719 "num_base_bdevs_discovered": 1, 00:08:38.719 "num_base_bdevs_operational": 2, 00:08:38.719 "base_bdevs_list": [ 00:08:38.719 { 00:08:38.719 "name": "BaseBdev1", 00:08:38.719 "uuid": "246a806b-c2c2-49ef-9243-13d444889ff1", 00:08:38.719 "is_configured": true, 00:08:38.719 "data_offset": 0, 00:08:38.719 "data_size": 65536 00:08:38.719 }, 00:08:38.719 { 00:08:38.719 "name": "BaseBdev2", 00:08:38.719 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:38.719 "is_configured": false, 00:08:38.719 "data_offset": 0, 00:08:38.719 "data_size": 0 00:08:38.719 } 00:08:38.719 ] 00:08:38.719 }' 00:08:38.719 15:14:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:38.719 15:14:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:38.979 15:14:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:38.979 15:14:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:38.979 15:14:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:38.979 [2024-11-19 15:14:29.212568] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:38.979 [2024-11-19 15:14:29.212660] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:08:38.979 15:14:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:38.979 15:14:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:08:38.979 15:14:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:38.979 15:14:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:38.979 [2024-11-19 15:14:29.224546] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:38.979 [2024-11-19 15:14:29.226747] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:38.979 [2024-11-19 15:14:29.226793] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:38.979 15:14:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:38.979 15:14:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:08:38.979 15:14:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:38.979 15:14:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:08:38.979 15:14:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:38.979 15:14:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:38.979 15:14:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:38.979 15:14:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:38.979 15:14:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:38.979 15:14:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:38.979 15:14:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:38.979 15:14:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:38.979 15:14:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:38.979 15:14:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:38.979 15:14:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:38.979 15:14:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:38.979 15:14:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:38.979 15:14:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:38.979 15:14:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:38.979 "name": "Existed_Raid", 00:08:38.979 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:38.979 "strip_size_kb": 64, 00:08:38.979 "state": "configuring", 00:08:38.979 "raid_level": "concat", 00:08:38.979 "superblock": false, 00:08:38.979 "num_base_bdevs": 2, 00:08:38.979 "num_base_bdevs_discovered": 1, 00:08:38.979 "num_base_bdevs_operational": 2, 00:08:38.979 "base_bdevs_list": [ 00:08:38.979 { 00:08:38.979 "name": "BaseBdev1", 00:08:38.979 "uuid": "246a806b-c2c2-49ef-9243-13d444889ff1", 00:08:38.979 "is_configured": true, 00:08:38.979 "data_offset": 0, 00:08:38.979 "data_size": 65536 00:08:38.979 }, 00:08:38.979 { 00:08:38.979 "name": "BaseBdev2", 00:08:38.979 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:38.979 "is_configured": false, 00:08:38.979 "data_offset": 0, 00:08:38.979 "data_size": 0 00:08:38.979 } 00:08:38.979 ] 00:08:38.979 }' 00:08:38.979 15:14:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:38.979 15:14:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:39.548 15:14:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:39.548 15:14:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:39.548 15:14:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:39.548 [2024-11-19 15:14:29.644747] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:39.548 [2024-11-19 15:14:29.644811] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:39.548 [2024-11-19 15:14:29.644820] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:08:39.548 [2024-11-19 15:14:29.645155] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:08:39.548 [2024-11-19 15:14:29.645361] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:39.548 [2024-11-19 15:14:29.645400] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:08:39.548 [2024-11-19 15:14:29.645637] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:39.548 BaseBdev2 00:08:39.548 15:14:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:39.548 15:14:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:08:39.548 15:14:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:08:39.548 15:14:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:39.548 15:14:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:39.548 15:14:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:39.548 15:14:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:39.548 15:14:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:39.548 15:14:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:39.548 15:14:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:39.548 15:14:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:39.548 15:14:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:39.548 15:14:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:39.548 15:14:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:39.548 [ 00:08:39.548 { 00:08:39.548 "name": "BaseBdev2", 00:08:39.548 "aliases": [ 00:08:39.548 "60803d2f-5285-45f7-b3fe-6a042878efb9" 00:08:39.548 ], 00:08:39.548 "product_name": "Malloc disk", 00:08:39.548 "block_size": 512, 00:08:39.548 "num_blocks": 65536, 00:08:39.548 "uuid": "60803d2f-5285-45f7-b3fe-6a042878efb9", 00:08:39.548 "assigned_rate_limits": { 00:08:39.548 "rw_ios_per_sec": 0, 00:08:39.548 "rw_mbytes_per_sec": 0, 00:08:39.548 "r_mbytes_per_sec": 0, 00:08:39.548 "w_mbytes_per_sec": 0 00:08:39.548 }, 00:08:39.548 "claimed": true, 00:08:39.548 "claim_type": "exclusive_write", 00:08:39.548 "zoned": false, 00:08:39.548 "supported_io_types": { 00:08:39.548 "read": true, 00:08:39.548 "write": true, 00:08:39.548 "unmap": true, 00:08:39.548 "flush": true, 00:08:39.548 "reset": true, 00:08:39.548 "nvme_admin": false, 00:08:39.548 "nvme_io": false, 00:08:39.548 "nvme_io_md": false, 00:08:39.548 "write_zeroes": true, 00:08:39.548 "zcopy": true, 00:08:39.548 "get_zone_info": false, 00:08:39.548 "zone_management": false, 00:08:39.548 "zone_append": false, 00:08:39.548 "compare": false, 00:08:39.548 "compare_and_write": false, 00:08:39.548 "abort": true, 00:08:39.548 "seek_hole": false, 00:08:39.548 "seek_data": false, 00:08:39.548 "copy": true, 00:08:39.548 "nvme_iov_md": false 00:08:39.548 }, 00:08:39.548 "memory_domains": [ 00:08:39.549 { 00:08:39.549 "dma_device_id": "system", 00:08:39.549 "dma_device_type": 1 00:08:39.549 }, 00:08:39.549 { 00:08:39.549 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:39.549 "dma_device_type": 2 00:08:39.549 } 00:08:39.549 ], 00:08:39.549 "driver_specific": {} 00:08:39.549 } 00:08:39.549 ] 00:08:39.549 15:14:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:39.549 15:14:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:39.549 15:14:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:39.549 15:14:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:39.549 15:14:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 2 00:08:39.549 15:14:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:39.549 15:14:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:39.549 15:14:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:39.549 15:14:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:39.549 15:14:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:39.549 15:14:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:39.549 15:14:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:39.549 15:14:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:39.549 15:14:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:39.549 15:14:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:39.549 15:14:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:39.549 15:14:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:39.549 15:14:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:39.549 15:14:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:39.549 15:14:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:39.549 "name": "Existed_Raid", 00:08:39.549 "uuid": "a352636f-6f28-4f63-a3e8-2be3be36e06c", 00:08:39.549 "strip_size_kb": 64, 00:08:39.549 "state": "online", 00:08:39.549 "raid_level": "concat", 00:08:39.549 "superblock": false, 00:08:39.549 "num_base_bdevs": 2, 00:08:39.549 "num_base_bdevs_discovered": 2, 00:08:39.549 "num_base_bdevs_operational": 2, 00:08:39.549 "base_bdevs_list": [ 00:08:39.549 { 00:08:39.549 "name": "BaseBdev1", 00:08:39.549 "uuid": "246a806b-c2c2-49ef-9243-13d444889ff1", 00:08:39.549 "is_configured": true, 00:08:39.549 "data_offset": 0, 00:08:39.549 "data_size": 65536 00:08:39.549 }, 00:08:39.549 { 00:08:39.549 "name": "BaseBdev2", 00:08:39.549 "uuid": "60803d2f-5285-45f7-b3fe-6a042878efb9", 00:08:39.549 "is_configured": true, 00:08:39.549 "data_offset": 0, 00:08:39.549 "data_size": 65536 00:08:39.549 } 00:08:39.549 ] 00:08:39.549 }' 00:08:39.549 15:14:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:39.549 15:14:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:39.808 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:08:39.808 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:39.808 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:39.808 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:39.808 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:39.808 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:39.808 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:39.808 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:39.808 15:14:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:39.808 15:14:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:39.808 [2024-11-19 15:14:30.040348] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:39.808 15:14:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:39.808 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:39.808 "name": "Existed_Raid", 00:08:39.808 "aliases": [ 00:08:39.808 "a352636f-6f28-4f63-a3e8-2be3be36e06c" 00:08:39.808 ], 00:08:39.808 "product_name": "Raid Volume", 00:08:39.808 "block_size": 512, 00:08:39.808 "num_blocks": 131072, 00:08:39.808 "uuid": "a352636f-6f28-4f63-a3e8-2be3be36e06c", 00:08:39.808 "assigned_rate_limits": { 00:08:39.808 "rw_ios_per_sec": 0, 00:08:39.808 "rw_mbytes_per_sec": 0, 00:08:39.808 "r_mbytes_per_sec": 0, 00:08:39.808 "w_mbytes_per_sec": 0 00:08:39.808 }, 00:08:39.808 "claimed": false, 00:08:39.808 "zoned": false, 00:08:39.808 "supported_io_types": { 00:08:39.808 "read": true, 00:08:39.808 "write": true, 00:08:39.808 "unmap": true, 00:08:39.808 "flush": true, 00:08:39.808 "reset": true, 00:08:39.808 "nvme_admin": false, 00:08:39.808 "nvme_io": false, 00:08:39.808 "nvme_io_md": false, 00:08:39.808 "write_zeroes": true, 00:08:39.808 "zcopy": false, 00:08:39.808 "get_zone_info": false, 00:08:39.808 "zone_management": false, 00:08:39.808 "zone_append": false, 00:08:39.808 "compare": false, 00:08:39.808 "compare_and_write": false, 00:08:39.808 "abort": false, 00:08:39.808 "seek_hole": false, 00:08:39.808 "seek_data": false, 00:08:39.808 "copy": false, 00:08:39.808 "nvme_iov_md": false 00:08:39.808 }, 00:08:39.808 "memory_domains": [ 00:08:39.808 { 00:08:39.808 "dma_device_id": "system", 00:08:39.808 "dma_device_type": 1 00:08:39.808 }, 00:08:39.808 { 00:08:39.808 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:39.808 "dma_device_type": 2 00:08:39.808 }, 00:08:39.808 { 00:08:39.808 "dma_device_id": "system", 00:08:39.808 "dma_device_type": 1 00:08:39.808 }, 00:08:39.808 { 00:08:39.808 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:39.808 "dma_device_type": 2 00:08:39.808 } 00:08:39.808 ], 00:08:39.808 "driver_specific": { 00:08:39.808 "raid": { 00:08:39.808 "uuid": "a352636f-6f28-4f63-a3e8-2be3be36e06c", 00:08:39.808 "strip_size_kb": 64, 00:08:39.808 "state": "online", 00:08:39.808 "raid_level": "concat", 00:08:39.808 "superblock": false, 00:08:39.808 "num_base_bdevs": 2, 00:08:39.808 "num_base_bdevs_discovered": 2, 00:08:39.808 "num_base_bdevs_operational": 2, 00:08:39.808 "base_bdevs_list": [ 00:08:39.808 { 00:08:39.808 "name": "BaseBdev1", 00:08:39.808 "uuid": "246a806b-c2c2-49ef-9243-13d444889ff1", 00:08:39.808 "is_configured": true, 00:08:39.808 "data_offset": 0, 00:08:39.808 "data_size": 65536 00:08:39.808 }, 00:08:39.808 { 00:08:39.808 "name": "BaseBdev2", 00:08:39.808 "uuid": "60803d2f-5285-45f7-b3fe-6a042878efb9", 00:08:39.808 "is_configured": true, 00:08:39.808 "data_offset": 0, 00:08:39.808 "data_size": 65536 00:08:39.808 } 00:08:39.808 ] 00:08:39.808 } 00:08:39.808 } 00:08:39.808 }' 00:08:39.808 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:39.808 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:08:39.808 BaseBdev2' 00:08:39.808 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:40.067 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:40.067 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:40.067 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:08:40.067 15:14:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.067 15:14:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.067 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:40.067 15:14:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.067 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:40.067 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:40.067 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:40.067 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:40.067 15:14:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.067 15:14:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.068 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:40.068 15:14:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.068 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:40.068 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:40.068 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:40.068 15:14:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.068 15:14:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.068 [2024-11-19 15:14:30.271808] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:40.068 [2024-11-19 15:14:30.271844] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:40.068 [2024-11-19 15:14:30.271896] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:40.068 15:14:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.068 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:08:40.068 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:08:40.068 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:40.068 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:40.068 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:08:40.068 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 1 00:08:40.068 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:40.068 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:08:40.068 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:40.068 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:40.068 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:08:40.068 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:40.068 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:40.068 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:40.068 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:40.068 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:40.068 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:40.068 15:14:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.068 15:14:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.068 15:14:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.068 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:40.068 "name": "Existed_Raid", 00:08:40.068 "uuid": "a352636f-6f28-4f63-a3e8-2be3be36e06c", 00:08:40.068 "strip_size_kb": 64, 00:08:40.068 "state": "offline", 00:08:40.068 "raid_level": "concat", 00:08:40.068 "superblock": false, 00:08:40.068 "num_base_bdevs": 2, 00:08:40.068 "num_base_bdevs_discovered": 1, 00:08:40.068 "num_base_bdevs_operational": 1, 00:08:40.068 "base_bdevs_list": [ 00:08:40.068 { 00:08:40.068 "name": null, 00:08:40.068 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:40.068 "is_configured": false, 00:08:40.068 "data_offset": 0, 00:08:40.068 "data_size": 65536 00:08:40.068 }, 00:08:40.068 { 00:08:40.068 "name": "BaseBdev2", 00:08:40.068 "uuid": "60803d2f-5285-45f7-b3fe-6a042878efb9", 00:08:40.068 "is_configured": true, 00:08:40.068 "data_offset": 0, 00:08:40.068 "data_size": 65536 00:08:40.068 } 00:08:40.068 ] 00:08:40.068 }' 00:08:40.068 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:40.068 15:14:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.638 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:08:40.638 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:40.638 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:40.638 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:40.638 15:14:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.638 15:14:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.638 15:14:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.638 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:40.638 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:40.638 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:08:40.638 15:14:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.638 15:14:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.638 [2024-11-19 15:14:30.731830] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:40.638 [2024-11-19 15:14:30.731912] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:08:40.638 15:14:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.638 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:40.638 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:40.638 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:08:40.638 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:40.638 15:14:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.638 15:14:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.638 15:14:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.638 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:08:40.638 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:08:40.638 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:08:40.638 15:14:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 73026 00:08:40.638 15:14:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 73026 ']' 00:08:40.638 15:14:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 73026 00:08:40.638 15:14:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:08:40.638 15:14:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:40.638 15:14:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 73026 00:08:40.638 15:14:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:40.638 15:14:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:40.638 killing process with pid 73026 00:08:40.638 15:14:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 73026' 00:08:40.638 15:14:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 73026 00:08:40.638 [2024-11-19 15:14:30.837222] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:40.638 15:14:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 73026 00:08:40.638 [2024-11-19 15:14:30.838783] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:40.898 15:14:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:08:40.898 00:08:40.898 real 0m3.804s 00:08:40.898 user 0m5.833s 00:08:40.898 sys 0m0.803s 00:08:40.898 15:14:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:40.898 15:14:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.898 ************************************ 00:08:40.898 END TEST raid_state_function_test 00:08:40.898 ************************************ 00:08:40.898 15:14:31 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test concat 2 true 00:08:40.898 15:14:31 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:40.898 15:14:31 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:40.898 15:14:31 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:40.898 ************************************ 00:08:40.898 START TEST raid_state_function_test_sb 00:08:40.898 ************************************ 00:08:40.898 15:14:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test concat 2 true 00:08:40.898 15:14:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:08:40.898 15:14:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:08:40.898 15:14:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:08:40.898 15:14:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:40.898 15:14:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:40.898 15:14:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:40.898 15:14:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:40.898 15:14:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:40.898 15:14:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:40.898 15:14:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:40.898 15:14:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:40.898 15:14:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:40.898 15:14:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:08:40.898 15:14:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:40.898 15:14:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:40.898 15:14:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:40.898 15:14:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:40.898 15:14:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:40.898 15:14:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:08:41.158 15:14:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:08:41.158 15:14:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:08:41.158 15:14:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:08:41.158 15:14:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:08:41.158 15:14:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=73268 00:08:41.158 15:14:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:41.158 Process raid pid: 73268 00:08:41.158 15:14:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 73268' 00:08:41.158 15:14:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 73268 00:08:41.158 15:14:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 73268 ']' 00:08:41.158 15:14:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:41.158 15:14:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:41.158 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:41.158 15:14:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:41.158 15:14:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:41.158 15:14:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:41.158 [2024-11-19 15:14:31.320282] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:08:41.158 [2024-11-19 15:14:31.320411] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:41.158 [2024-11-19 15:14:31.477383] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:41.418 [2024-11-19 15:14:31.515609] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:41.418 [2024-11-19 15:14:31.591510] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:41.418 [2024-11-19 15:14:31.591557] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:41.987 15:14:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:41.987 15:14:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:08:41.987 15:14:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:08:41.987 15:14:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:41.987 15:14:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:41.987 [2024-11-19 15:14:32.166949] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:41.987 [2024-11-19 15:14:32.167022] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:41.987 [2024-11-19 15:14:32.167060] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:41.987 [2024-11-19 15:14:32.167071] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:41.987 15:14:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:41.987 15:14:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:08:41.987 15:14:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:41.987 15:14:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:41.987 15:14:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:41.987 15:14:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:41.987 15:14:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:41.987 15:14:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:41.987 15:14:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:41.987 15:14:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:41.987 15:14:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:41.987 15:14:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:41.987 15:14:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:41.987 15:14:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:41.987 15:14:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:41.987 15:14:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:41.987 15:14:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:41.987 "name": "Existed_Raid", 00:08:41.987 "uuid": "a7b0b05f-35e8-4fff-b04a-6c253313b03d", 00:08:41.987 "strip_size_kb": 64, 00:08:41.987 "state": "configuring", 00:08:41.987 "raid_level": "concat", 00:08:41.987 "superblock": true, 00:08:41.987 "num_base_bdevs": 2, 00:08:41.987 "num_base_bdevs_discovered": 0, 00:08:41.987 "num_base_bdevs_operational": 2, 00:08:41.987 "base_bdevs_list": [ 00:08:41.987 { 00:08:41.987 "name": "BaseBdev1", 00:08:41.987 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:41.987 "is_configured": false, 00:08:41.987 "data_offset": 0, 00:08:41.987 "data_size": 0 00:08:41.987 }, 00:08:41.987 { 00:08:41.987 "name": "BaseBdev2", 00:08:41.987 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:41.987 "is_configured": false, 00:08:41.987 "data_offset": 0, 00:08:41.987 "data_size": 0 00:08:41.987 } 00:08:41.987 ] 00:08:41.987 }' 00:08:41.987 15:14:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:41.987 15:14:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:42.576 15:14:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:42.576 15:14:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:42.576 15:14:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:42.576 [2024-11-19 15:14:32.602142] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:42.576 [2024-11-19 15:14:32.602214] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:08:42.576 15:14:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:42.576 15:14:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:08:42.576 15:14:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:42.576 15:14:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:42.576 [2024-11-19 15:14:32.610100] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:42.576 [2024-11-19 15:14:32.610147] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:42.576 [2024-11-19 15:14:32.610157] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:42.576 [2024-11-19 15:14:32.610180] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:42.576 15:14:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:42.576 15:14:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:42.576 15:14:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:42.576 15:14:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:42.576 [2024-11-19 15:14:32.633271] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:42.576 BaseBdev1 00:08:42.576 15:14:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:42.576 15:14:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:08:42.576 15:14:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:08:42.576 15:14:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:42.576 15:14:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:42.576 15:14:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:42.576 15:14:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:42.576 15:14:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:42.576 15:14:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:42.576 15:14:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:42.576 15:14:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:42.576 15:14:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:42.576 15:14:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:42.576 15:14:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:42.576 [ 00:08:42.576 { 00:08:42.576 "name": "BaseBdev1", 00:08:42.576 "aliases": [ 00:08:42.576 "215bbd05-d1cd-4898-a2b3-8a70d7f4d9d6" 00:08:42.576 ], 00:08:42.576 "product_name": "Malloc disk", 00:08:42.576 "block_size": 512, 00:08:42.576 "num_blocks": 65536, 00:08:42.576 "uuid": "215bbd05-d1cd-4898-a2b3-8a70d7f4d9d6", 00:08:42.576 "assigned_rate_limits": { 00:08:42.576 "rw_ios_per_sec": 0, 00:08:42.576 "rw_mbytes_per_sec": 0, 00:08:42.576 "r_mbytes_per_sec": 0, 00:08:42.576 "w_mbytes_per_sec": 0 00:08:42.576 }, 00:08:42.576 "claimed": true, 00:08:42.576 "claim_type": "exclusive_write", 00:08:42.576 "zoned": false, 00:08:42.576 "supported_io_types": { 00:08:42.576 "read": true, 00:08:42.576 "write": true, 00:08:42.576 "unmap": true, 00:08:42.576 "flush": true, 00:08:42.576 "reset": true, 00:08:42.576 "nvme_admin": false, 00:08:42.576 "nvme_io": false, 00:08:42.576 "nvme_io_md": false, 00:08:42.576 "write_zeroes": true, 00:08:42.576 "zcopy": true, 00:08:42.576 "get_zone_info": false, 00:08:42.576 "zone_management": false, 00:08:42.576 "zone_append": false, 00:08:42.576 "compare": false, 00:08:42.576 "compare_and_write": false, 00:08:42.576 "abort": true, 00:08:42.576 "seek_hole": false, 00:08:42.576 "seek_data": false, 00:08:42.576 "copy": true, 00:08:42.576 "nvme_iov_md": false 00:08:42.576 }, 00:08:42.576 "memory_domains": [ 00:08:42.576 { 00:08:42.576 "dma_device_id": "system", 00:08:42.576 "dma_device_type": 1 00:08:42.576 }, 00:08:42.576 { 00:08:42.576 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:42.576 "dma_device_type": 2 00:08:42.576 } 00:08:42.576 ], 00:08:42.576 "driver_specific": {} 00:08:42.576 } 00:08:42.576 ] 00:08:42.576 15:14:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:42.576 15:14:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:42.576 15:14:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:08:42.576 15:14:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:42.576 15:14:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:42.576 15:14:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:42.576 15:14:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:42.576 15:14:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:42.576 15:14:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:42.576 15:14:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:42.576 15:14:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:42.576 15:14:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:42.576 15:14:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:42.576 15:14:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:42.576 15:14:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:42.576 15:14:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:42.576 15:14:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:42.576 15:14:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:42.576 "name": "Existed_Raid", 00:08:42.576 "uuid": "1c175955-784d-4efd-97db-bf9a4eb2e76e", 00:08:42.576 "strip_size_kb": 64, 00:08:42.576 "state": "configuring", 00:08:42.576 "raid_level": "concat", 00:08:42.576 "superblock": true, 00:08:42.576 "num_base_bdevs": 2, 00:08:42.576 "num_base_bdevs_discovered": 1, 00:08:42.576 "num_base_bdevs_operational": 2, 00:08:42.577 "base_bdevs_list": [ 00:08:42.577 { 00:08:42.577 "name": "BaseBdev1", 00:08:42.577 "uuid": "215bbd05-d1cd-4898-a2b3-8a70d7f4d9d6", 00:08:42.577 "is_configured": true, 00:08:42.577 "data_offset": 2048, 00:08:42.577 "data_size": 63488 00:08:42.577 }, 00:08:42.577 { 00:08:42.577 "name": "BaseBdev2", 00:08:42.577 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:42.577 "is_configured": false, 00:08:42.577 "data_offset": 0, 00:08:42.577 "data_size": 0 00:08:42.577 } 00:08:42.577 ] 00:08:42.577 }' 00:08:42.577 15:14:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:42.577 15:14:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:42.837 15:14:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:42.837 15:14:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:42.837 15:14:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:42.837 [2024-11-19 15:14:33.080531] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:42.837 [2024-11-19 15:14:33.080585] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:08:42.837 15:14:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:42.837 15:14:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:08:42.837 15:14:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:42.837 15:14:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:42.837 [2024-11-19 15:14:33.092549] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:42.837 [2024-11-19 15:14:33.094709] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:42.837 [2024-11-19 15:14:33.094769] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:42.837 15:14:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:42.837 15:14:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:08:42.837 15:14:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:42.837 15:14:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:08:42.837 15:14:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:42.837 15:14:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:42.837 15:14:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:42.837 15:14:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:42.837 15:14:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:42.837 15:14:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:42.837 15:14:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:42.837 15:14:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:42.837 15:14:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:42.837 15:14:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:42.837 15:14:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:42.837 15:14:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:42.837 15:14:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:42.837 15:14:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:42.837 15:14:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:42.837 "name": "Existed_Raid", 00:08:42.837 "uuid": "07966db6-7bc5-479e-b6c7-37f0d9d6988a", 00:08:42.837 "strip_size_kb": 64, 00:08:42.837 "state": "configuring", 00:08:42.837 "raid_level": "concat", 00:08:42.837 "superblock": true, 00:08:42.837 "num_base_bdevs": 2, 00:08:42.837 "num_base_bdevs_discovered": 1, 00:08:42.837 "num_base_bdevs_operational": 2, 00:08:42.837 "base_bdevs_list": [ 00:08:42.837 { 00:08:42.837 "name": "BaseBdev1", 00:08:42.837 "uuid": "215bbd05-d1cd-4898-a2b3-8a70d7f4d9d6", 00:08:42.837 "is_configured": true, 00:08:42.837 "data_offset": 2048, 00:08:42.837 "data_size": 63488 00:08:42.837 }, 00:08:42.837 { 00:08:42.837 "name": "BaseBdev2", 00:08:42.837 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:42.837 "is_configured": false, 00:08:42.837 "data_offset": 0, 00:08:42.837 "data_size": 0 00:08:42.837 } 00:08:42.837 ] 00:08:42.837 }' 00:08:42.837 15:14:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:42.837 15:14:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:43.406 15:14:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:43.406 15:14:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.406 15:14:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:43.406 [2024-11-19 15:14:33.520910] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:43.406 [2024-11-19 15:14:33.521177] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:43.406 [2024-11-19 15:14:33.521195] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:08:43.406 [2024-11-19 15:14:33.521542] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:08:43.406 BaseBdev2 00:08:43.406 [2024-11-19 15:14:33.521747] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:43.406 [2024-11-19 15:14:33.521780] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:08:43.406 [2024-11-19 15:14:33.521943] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:43.406 15:14:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.406 15:14:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:08:43.407 15:14:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:08:43.407 15:14:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:43.407 15:14:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:43.407 15:14:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:43.407 15:14:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:43.407 15:14:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:43.407 15:14:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.407 15:14:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:43.407 15:14:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.407 15:14:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:43.407 15:14:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.407 15:14:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:43.407 [ 00:08:43.407 { 00:08:43.407 "name": "BaseBdev2", 00:08:43.407 "aliases": [ 00:08:43.407 "b42a1761-8c82-4786-a43c-c614ac46b279" 00:08:43.407 ], 00:08:43.407 "product_name": "Malloc disk", 00:08:43.407 "block_size": 512, 00:08:43.407 "num_blocks": 65536, 00:08:43.407 "uuid": "b42a1761-8c82-4786-a43c-c614ac46b279", 00:08:43.407 "assigned_rate_limits": { 00:08:43.407 "rw_ios_per_sec": 0, 00:08:43.407 "rw_mbytes_per_sec": 0, 00:08:43.407 "r_mbytes_per_sec": 0, 00:08:43.407 "w_mbytes_per_sec": 0 00:08:43.407 }, 00:08:43.407 "claimed": true, 00:08:43.407 "claim_type": "exclusive_write", 00:08:43.407 "zoned": false, 00:08:43.407 "supported_io_types": { 00:08:43.407 "read": true, 00:08:43.407 "write": true, 00:08:43.407 "unmap": true, 00:08:43.407 "flush": true, 00:08:43.407 "reset": true, 00:08:43.407 "nvme_admin": false, 00:08:43.407 "nvme_io": false, 00:08:43.407 "nvme_io_md": false, 00:08:43.407 "write_zeroes": true, 00:08:43.407 "zcopy": true, 00:08:43.407 "get_zone_info": false, 00:08:43.407 "zone_management": false, 00:08:43.407 "zone_append": false, 00:08:43.407 "compare": false, 00:08:43.407 "compare_and_write": false, 00:08:43.407 "abort": true, 00:08:43.407 "seek_hole": false, 00:08:43.407 "seek_data": false, 00:08:43.407 "copy": true, 00:08:43.407 "nvme_iov_md": false 00:08:43.407 }, 00:08:43.407 "memory_domains": [ 00:08:43.407 { 00:08:43.407 "dma_device_id": "system", 00:08:43.407 "dma_device_type": 1 00:08:43.407 }, 00:08:43.407 { 00:08:43.407 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:43.407 "dma_device_type": 2 00:08:43.407 } 00:08:43.407 ], 00:08:43.407 "driver_specific": {} 00:08:43.407 } 00:08:43.407 ] 00:08:43.407 15:14:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.407 15:14:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:43.407 15:14:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:43.407 15:14:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:43.407 15:14:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 2 00:08:43.407 15:14:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:43.407 15:14:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:43.407 15:14:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:43.407 15:14:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:43.407 15:14:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:43.407 15:14:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:43.407 15:14:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:43.407 15:14:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:43.407 15:14:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:43.407 15:14:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:43.407 15:14:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:43.407 15:14:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.407 15:14:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:43.407 15:14:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.407 15:14:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:43.407 "name": "Existed_Raid", 00:08:43.407 "uuid": "07966db6-7bc5-479e-b6c7-37f0d9d6988a", 00:08:43.407 "strip_size_kb": 64, 00:08:43.407 "state": "online", 00:08:43.407 "raid_level": "concat", 00:08:43.407 "superblock": true, 00:08:43.407 "num_base_bdevs": 2, 00:08:43.407 "num_base_bdevs_discovered": 2, 00:08:43.407 "num_base_bdevs_operational": 2, 00:08:43.407 "base_bdevs_list": [ 00:08:43.407 { 00:08:43.407 "name": "BaseBdev1", 00:08:43.407 "uuid": "215bbd05-d1cd-4898-a2b3-8a70d7f4d9d6", 00:08:43.407 "is_configured": true, 00:08:43.407 "data_offset": 2048, 00:08:43.407 "data_size": 63488 00:08:43.407 }, 00:08:43.407 { 00:08:43.407 "name": "BaseBdev2", 00:08:43.407 "uuid": "b42a1761-8c82-4786-a43c-c614ac46b279", 00:08:43.407 "is_configured": true, 00:08:43.407 "data_offset": 2048, 00:08:43.407 "data_size": 63488 00:08:43.407 } 00:08:43.407 ] 00:08:43.407 }' 00:08:43.407 15:14:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:43.407 15:14:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:43.976 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:08:43.976 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:43.976 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:43.976 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:43.976 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:08:43.976 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:43.976 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:43.976 15:14:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.976 15:14:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:43.976 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:43.976 [2024-11-19 15:14:34.020425] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:43.976 15:14:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.976 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:43.976 "name": "Existed_Raid", 00:08:43.976 "aliases": [ 00:08:43.976 "07966db6-7bc5-479e-b6c7-37f0d9d6988a" 00:08:43.976 ], 00:08:43.976 "product_name": "Raid Volume", 00:08:43.976 "block_size": 512, 00:08:43.976 "num_blocks": 126976, 00:08:43.976 "uuid": "07966db6-7bc5-479e-b6c7-37f0d9d6988a", 00:08:43.976 "assigned_rate_limits": { 00:08:43.976 "rw_ios_per_sec": 0, 00:08:43.976 "rw_mbytes_per_sec": 0, 00:08:43.976 "r_mbytes_per_sec": 0, 00:08:43.976 "w_mbytes_per_sec": 0 00:08:43.976 }, 00:08:43.976 "claimed": false, 00:08:43.976 "zoned": false, 00:08:43.976 "supported_io_types": { 00:08:43.976 "read": true, 00:08:43.976 "write": true, 00:08:43.976 "unmap": true, 00:08:43.976 "flush": true, 00:08:43.976 "reset": true, 00:08:43.976 "nvme_admin": false, 00:08:43.976 "nvme_io": false, 00:08:43.976 "nvme_io_md": false, 00:08:43.976 "write_zeroes": true, 00:08:43.976 "zcopy": false, 00:08:43.976 "get_zone_info": false, 00:08:43.976 "zone_management": false, 00:08:43.976 "zone_append": false, 00:08:43.976 "compare": false, 00:08:43.976 "compare_and_write": false, 00:08:43.976 "abort": false, 00:08:43.976 "seek_hole": false, 00:08:43.976 "seek_data": false, 00:08:43.976 "copy": false, 00:08:43.976 "nvme_iov_md": false 00:08:43.976 }, 00:08:43.976 "memory_domains": [ 00:08:43.976 { 00:08:43.976 "dma_device_id": "system", 00:08:43.976 "dma_device_type": 1 00:08:43.976 }, 00:08:43.976 { 00:08:43.976 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:43.976 "dma_device_type": 2 00:08:43.976 }, 00:08:43.976 { 00:08:43.976 "dma_device_id": "system", 00:08:43.976 "dma_device_type": 1 00:08:43.976 }, 00:08:43.976 { 00:08:43.976 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:43.976 "dma_device_type": 2 00:08:43.976 } 00:08:43.976 ], 00:08:43.976 "driver_specific": { 00:08:43.976 "raid": { 00:08:43.976 "uuid": "07966db6-7bc5-479e-b6c7-37f0d9d6988a", 00:08:43.976 "strip_size_kb": 64, 00:08:43.976 "state": "online", 00:08:43.976 "raid_level": "concat", 00:08:43.976 "superblock": true, 00:08:43.976 "num_base_bdevs": 2, 00:08:43.976 "num_base_bdevs_discovered": 2, 00:08:43.976 "num_base_bdevs_operational": 2, 00:08:43.976 "base_bdevs_list": [ 00:08:43.976 { 00:08:43.976 "name": "BaseBdev1", 00:08:43.976 "uuid": "215bbd05-d1cd-4898-a2b3-8a70d7f4d9d6", 00:08:43.976 "is_configured": true, 00:08:43.976 "data_offset": 2048, 00:08:43.976 "data_size": 63488 00:08:43.976 }, 00:08:43.976 { 00:08:43.976 "name": "BaseBdev2", 00:08:43.976 "uuid": "b42a1761-8c82-4786-a43c-c614ac46b279", 00:08:43.976 "is_configured": true, 00:08:43.976 "data_offset": 2048, 00:08:43.976 "data_size": 63488 00:08:43.976 } 00:08:43.976 ] 00:08:43.976 } 00:08:43.976 } 00:08:43.976 }' 00:08:43.976 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:43.976 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:08:43.976 BaseBdev2' 00:08:43.976 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:43.976 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:43.976 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:43.976 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:08:43.976 15:14:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.976 15:14:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:43.976 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:43.976 15:14:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.977 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:43.977 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:43.977 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:43.977 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:43.977 15:14:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.977 15:14:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:43.977 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:43.977 15:14:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.977 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:43.977 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:43.977 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:43.977 15:14:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.977 15:14:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:43.977 [2024-11-19 15:14:34.267872] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:43.977 [2024-11-19 15:14:34.267906] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:43.977 [2024-11-19 15:14:34.267991] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:43.977 15:14:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.977 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:08:43.977 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:08:43.977 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:43.977 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:08:43.977 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:08:43.977 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 1 00:08:43.977 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:43.977 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:08:43.977 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:43.977 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:43.977 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:08:43.977 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:43.977 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:43.977 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:43.977 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:43.977 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:43.977 15:14:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.977 15:14:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:43.977 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:43.977 15:14:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:44.237 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:44.237 "name": "Existed_Raid", 00:08:44.237 "uuid": "07966db6-7bc5-479e-b6c7-37f0d9d6988a", 00:08:44.237 "strip_size_kb": 64, 00:08:44.237 "state": "offline", 00:08:44.237 "raid_level": "concat", 00:08:44.237 "superblock": true, 00:08:44.237 "num_base_bdevs": 2, 00:08:44.237 "num_base_bdevs_discovered": 1, 00:08:44.237 "num_base_bdevs_operational": 1, 00:08:44.237 "base_bdevs_list": [ 00:08:44.237 { 00:08:44.237 "name": null, 00:08:44.237 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:44.237 "is_configured": false, 00:08:44.237 "data_offset": 0, 00:08:44.237 "data_size": 63488 00:08:44.237 }, 00:08:44.237 { 00:08:44.237 "name": "BaseBdev2", 00:08:44.237 "uuid": "b42a1761-8c82-4786-a43c-c614ac46b279", 00:08:44.237 "is_configured": true, 00:08:44.237 "data_offset": 2048, 00:08:44.237 "data_size": 63488 00:08:44.237 } 00:08:44.237 ] 00:08:44.237 }' 00:08:44.237 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:44.237 15:14:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:44.497 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:08:44.497 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:44.497 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:44.497 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:44.497 15:14:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:44.497 15:14:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:44.497 15:14:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:44.497 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:44.497 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:44.497 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:08:44.497 15:14:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:44.497 15:14:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:44.497 [2024-11-19 15:14:34.799912] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:44.497 [2024-11-19 15:14:34.799996] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:08:44.497 15:14:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:44.497 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:44.497 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:44.497 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:08:44.497 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:44.497 15:14:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:44.497 15:14:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:44.757 15:14:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:44.757 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:08:44.757 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:08:44.757 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:08:44.757 15:14:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 73268 00:08:44.757 15:14:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 73268 ']' 00:08:44.757 15:14:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 73268 00:08:44.757 15:14:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:08:44.757 15:14:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:44.757 15:14:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 73268 00:08:44.757 15:14:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:44.757 15:14:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:44.757 killing process with pid 73268 00:08:44.757 15:14:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 73268' 00:08:44.757 15:14:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 73268 00:08:44.757 [2024-11-19 15:14:34.904205] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:44.757 15:14:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 73268 00:08:44.757 [2024-11-19 15:14:34.905765] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:45.017 15:14:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:08:45.017 00:08:45.017 real 0m4.001s 00:08:45.017 user 0m6.193s 00:08:45.017 sys 0m0.811s 00:08:45.017 15:14:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:45.017 15:14:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:45.017 ************************************ 00:08:45.017 END TEST raid_state_function_test_sb 00:08:45.017 ************************************ 00:08:45.017 15:14:35 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test concat 2 00:08:45.017 15:14:35 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:08:45.017 15:14:35 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:45.017 15:14:35 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:45.017 ************************************ 00:08:45.017 START TEST raid_superblock_test 00:08:45.017 ************************************ 00:08:45.017 15:14:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test concat 2 00:08:45.017 15:14:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=concat 00:08:45.017 15:14:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:08:45.017 15:14:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:08:45.017 15:14:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:08:45.017 15:14:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:08:45.017 15:14:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:08:45.017 15:14:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:08:45.017 15:14:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:08:45.017 15:14:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:08:45.017 15:14:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:08:45.017 15:14:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:08:45.017 15:14:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:08:45.017 15:14:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:08:45.017 15:14:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' concat '!=' raid1 ']' 00:08:45.017 15:14:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:08:45.017 15:14:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:08:45.017 15:14:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=73504 00:08:45.017 15:14:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:08:45.017 15:14:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 73504 00:08:45.017 15:14:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 73504 ']' 00:08:45.017 15:14:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:45.017 15:14:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:45.017 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:45.017 15:14:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:45.017 15:14:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:45.017 15:14:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.277 [2024-11-19 15:14:35.387748] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:08:45.277 [2024-11-19 15:14:35.387865] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid73504 ] 00:08:45.277 [2024-11-19 15:14:35.522995] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:45.277 [2024-11-19 15:14:35.561206] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:45.537 [2024-11-19 15:14:35.639814] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:45.537 [2024-11-19 15:14:35.639872] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:46.106 15:14:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:46.106 15:14:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:08:46.106 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:08:46.106 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:46.106 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:08:46.106 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:08:46.106 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:08:46.106 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:46.106 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:46.106 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:46.106 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:08:46.106 15:14:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.106 15:14:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.106 malloc1 00:08:46.106 15:14:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.106 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:46.106 15:14:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.106 15:14:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.106 [2024-11-19 15:14:36.254094] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:46.106 [2024-11-19 15:14:36.254168] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:46.106 [2024-11-19 15:14:36.254188] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:08:46.106 [2024-11-19 15:14:36.254210] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:46.106 [2024-11-19 15:14:36.256724] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:46.106 [2024-11-19 15:14:36.256769] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:46.106 pt1 00:08:46.106 15:14:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.106 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:46.106 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:46.106 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:08:46.106 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:08:46.106 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:08:46.106 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:46.106 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:46.106 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:46.106 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:08:46.106 15:14:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.106 15:14:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.106 malloc2 00:08:46.106 15:14:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.106 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:46.106 15:14:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.106 15:14:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.106 [2024-11-19 15:14:36.288526] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:46.106 [2024-11-19 15:14:36.288595] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:46.106 [2024-11-19 15:14:36.288611] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:08:46.106 [2024-11-19 15:14:36.288622] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:46.106 [2024-11-19 15:14:36.290999] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:46.106 [2024-11-19 15:14:36.291034] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:46.106 pt2 00:08:46.106 15:14:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.106 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:46.106 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:46.106 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:08:46.106 15:14:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.106 15:14:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.106 [2024-11-19 15:14:36.300561] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:46.106 [2024-11-19 15:14:36.302741] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:46.106 [2024-11-19 15:14:36.302887] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:08:46.107 [2024-11-19 15:14:36.302902] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:08:46.107 [2024-11-19 15:14:36.303230] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:08:46.107 [2024-11-19 15:14:36.303389] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:08:46.107 [2024-11-19 15:14:36.303399] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:08:46.107 [2024-11-19 15:14:36.303534] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:46.107 15:14:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.107 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:08:46.107 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:46.107 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:46.107 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:46.107 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:46.107 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:46.107 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:46.107 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:46.107 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:46.107 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:46.107 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:46.107 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:46.107 15:14:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.107 15:14:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.107 15:14:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.107 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:46.107 "name": "raid_bdev1", 00:08:46.107 "uuid": "f5025517-ed2d-46b1-9c3e-d63b904bb272", 00:08:46.107 "strip_size_kb": 64, 00:08:46.107 "state": "online", 00:08:46.107 "raid_level": "concat", 00:08:46.107 "superblock": true, 00:08:46.107 "num_base_bdevs": 2, 00:08:46.107 "num_base_bdevs_discovered": 2, 00:08:46.107 "num_base_bdevs_operational": 2, 00:08:46.107 "base_bdevs_list": [ 00:08:46.107 { 00:08:46.107 "name": "pt1", 00:08:46.107 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:46.107 "is_configured": true, 00:08:46.107 "data_offset": 2048, 00:08:46.107 "data_size": 63488 00:08:46.107 }, 00:08:46.107 { 00:08:46.107 "name": "pt2", 00:08:46.107 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:46.107 "is_configured": true, 00:08:46.107 "data_offset": 2048, 00:08:46.107 "data_size": 63488 00:08:46.107 } 00:08:46.107 ] 00:08:46.107 }' 00:08:46.107 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:46.107 15:14:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.674 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:08:46.674 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:08:46.674 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:46.674 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:46.674 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:46.674 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:46.674 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:46.674 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:46.674 15:14:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.674 15:14:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.674 [2024-11-19 15:14:36.768138] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:46.674 15:14:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.674 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:46.674 "name": "raid_bdev1", 00:08:46.674 "aliases": [ 00:08:46.674 "f5025517-ed2d-46b1-9c3e-d63b904bb272" 00:08:46.674 ], 00:08:46.674 "product_name": "Raid Volume", 00:08:46.674 "block_size": 512, 00:08:46.674 "num_blocks": 126976, 00:08:46.674 "uuid": "f5025517-ed2d-46b1-9c3e-d63b904bb272", 00:08:46.674 "assigned_rate_limits": { 00:08:46.674 "rw_ios_per_sec": 0, 00:08:46.674 "rw_mbytes_per_sec": 0, 00:08:46.674 "r_mbytes_per_sec": 0, 00:08:46.674 "w_mbytes_per_sec": 0 00:08:46.674 }, 00:08:46.674 "claimed": false, 00:08:46.674 "zoned": false, 00:08:46.674 "supported_io_types": { 00:08:46.674 "read": true, 00:08:46.674 "write": true, 00:08:46.674 "unmap": true, 00:08:46.674 "flush": true, 00:08:46.674 "reset": true, 00:08:46.674 "nvme_admin": false, 00:08:46.674 "nvme_io": false, 00:08:46.674 "nvme_io_md": false, 00:08:46.674 "write_zeroes": true, 00:08:46.674 "zcopy": false, 00:08:46.674 "get_zone_info": false, 00:08:46.674 "zone_management": false, 00:08:46.674 "zone_append": false, 00:08:46.674 "compare": false, 00:08:46.674 "compare_and_write": false, 00:08:46.675 "abort": false, 00:08:46.675 "seek_hole": false, 00:08:46.675 "seek_data": false, 00:08:46.675 "copy": false, 00:08:46.675 "nvme_iov_md": false 00:08:46.675 }, 00:08:46.675 "memory_domains": [ 00:08:46.675 { 00:08:46.675 "dma_device_id": "system", 00:08:46.675 "dma_device_type": 1 00:08:46.675 }, 00:08:46.675 { 00:08:46.675 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:46.675 "dma_device_type": 2 00:08:46.675 }, 00:08:46.675 { 00:08:46.675 "dma_device_id": "system", 00:08:46.675 "dma_device_type": 1 00:08:46.675 }, 00:08:46.675 { 00:08:46.675 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:46.675 "dma_device_type": 2 00:08:46.675 } 00:08:46.675 ], 00:08:46.675 "driver_specific": { 00:08:46.675 "raid": { 00:08:46.675 "uuid": "f5025517-ed2d-46b1-9c3e-d63b904bb272", 00:08:46.675 "strip_size_kb": 64, 00:08:46.675 "state": "online", 00:08:46.675 "raid_level": "concat", 00:08:46.675 "superblock": true, 00:08:46.675 "num_base_bdevs": 2, 00:08:46.675 "num_base_bdevs_discovered": 2, 00:08:46.675 "num_base_bdevs_operational": 2, 00:08:46.675 "base_bdevs_list": [ 00:08:46.675 { 00:08:46.675 "name": "pt1", 00:08:46.675 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:46.675 "is_configured": true, 00:08:46.675 "data_offset": 2048, 00:08:46.675 "data_size": 63488 00:08:46.675 }, 00:08:46.675 { 00:08:46.675 "name": "pt2", 00:08:46.675 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:46.675 "is_configured": true, 00:08:46.675 "data_offset": 2048, 00:08:46.675 "data_size": 63488 00:08:46.675 } 00:08:46.675 ] 00:08:46.675 } 00:08:46.675 } 00:08:46.675 }' 00:08:46.675 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:46.675 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:08:46.675 pt2' 00:08:46.675 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:46.675 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:46.675 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:46.675 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:08:46.675 15:14:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.675 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:46.675 15:14:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.675 15:14:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.675 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:46.675 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:46.675 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:46.675 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:08:46.675 15:14:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.675 15:14:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.675 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:46.675 15:14:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.675 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:46.675 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:46.675 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:46.675 15:14:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.675 15:14:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.675 15:14:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:08:46.675 [2024-11-19 15:14:36.971608] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:46.675 15:14:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.675 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=f5025517-ed2d-46b1-9c3e-d63b904bb272 00:08:46.675 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z f5025517-ed2d-46b1-9c3e-d63b904bb272 ']' 00:08:46.675 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:46.675 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.675 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.934 [2024-11-19 15:14:37.015314] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:46.934 [2024-11-19 15:14:37.015351] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:46.934 [2024-11-19 15:14:37.015422] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:46.934 [2024-11-19 15:14:37.015472] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:46.934 [2024-11-19 15:14:37.015485] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.934 [2024-11-19 15:14:37.147124] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:08:46.934 [2024-11-19 15:14:37.149487] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:08:46.934 [2024-11-19 15:14:37.149554] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:08:46.934 [2024-11-19 15:14:37.149598] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:08:46.934 [2024-11-19 15:14:37.149614] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:46.934 [2024-11-19 15:14:37.149624] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:08:46.934 request: 00:08:46.934 { 00:08:46.934 "name": "raid_bdev1", 00:08:46.934 "raid_level": "concat", 00:08:46.934 "base_bdevs": [ 00:08:46.934 "malloc1", 00:08:46.934 "malloc2" 00:08:46.934 ], 00:08:46.934 "strip_size_kb": 64, 00:08:46.934 "superblock": false, 00:08:46.934 "method": "bdev_raid_create", 00:08:46.934 "req_id": 1 00:08:46.934 } 00:08:46.934 Got JSON-RPC error response 00:08:46.934 response: 00:08:46.934 { 00:08:46.934 "code": -17, 00:08:46.934 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:08:46.934 } 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.934 [2024-11-19 15:14:37.210998] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:46.934 [2024-11-19 15:14:37.211046] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:46.934 [2024-11-19 15:14:37.211076] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:08:46.934 [2024-11-19 15:14:37.211085] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:46.934 [2024-11-19 15:14:37.213656] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:46.934 [2024-11-19 15:14:37.213690] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:46.934 [2024-11-19 15:14:37.213755] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:08:46.934 [2024-11-19 15:14:37.213799] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:46.934 pt1 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 2 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:46.934 "name": "raid_bdev1", 00:08:46.934 "uuid": "f5025517-ed2d-46b1-9c3e-d63b904bb272", 00:08:46.934 "strip_size_kb": 64, 00:08:46.934 "state": "configuring", 00:08:46.934 "raid_level": "concat", 00:08:46.934 "superblock": true, 00:08:46.934 "num_base_bdevs": 2, 00:08:46.934 "num_base_bdevs_discovered": 1, 00:08:46.934 "num_base_bdevs_operational": 2, 00:08:46.934 "base_bdevs_list": [ 00:08:46.934 { 00:08:46.934 "name": "pt1", 00:08:46.934 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:46.934 "is_configured": true, 00:08:46.934 "data_offset": 2048, 00:08:46.934 "data_size": 63488 00:08:46.934 }, 00:08:46.934 { 00:08:46.934 "name": null, 00:08:46.934 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:46.934 "is_configured": false, 00:08:46.934 "data_offset": 2048, 00:08:46.934 "data_size": 63488 00:08:46.934 } 00:08:46.934 ] 00:08:46.934 }' 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:46.934 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.503 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:08:47.503 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:08:47.503 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:47.503 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:47.503 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.503 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.503 [2024-11-19 15:14:37.666176] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:47.503 [2024-11-19 15:14:37.666290] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:47.503 [2024-11-19 15:14:37.666327] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:08:47.503 [2024-11-19 15:14:37.666354] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:47.503 [2024-11-19 15:14:37.666784] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:47.503 [2024-11-19 15:14:37.666841] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:47.503 [2024-11-19 15:14:37.666934] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:08:47.503 [2024-11-19 15:14:37.666989] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:47.503 [2024-11-19 15:14:37.667102] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:47.503 [2024-11-19 15:14:37.667141] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:08:47.503 [2024-11-19 15:14:37.667434] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:08:47.503 [2024-11-19 15:14:37.667593] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:47.503 [2024-11-19 15:14:37.667640] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:08:47.503 [2024-11-19 15:14:37.667783] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:47.503 pt2 00:08:47.503 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.503 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:08:47.503 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:47.503 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:08:47.503 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:47.503 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:47.503 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:47.503 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:47.503 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:47.503 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:47.503 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:47.503 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:47.503 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:47.503 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:47.503 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:47.503 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.503 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.503 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.503 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:47.503 "name": "raid_bdev1", 00:08:47.503 "uuid": "f5025517-ed2d-46b1-9c3e-d63b904bb272", 00:08:47.503 "strip_size_kb": 64, 00:08:47.503 "state": "online", 00:08:47.503 "raid_level": "concat", 00:08:47.503 "superblock": true, 00:08:47.503 "num_base_bdevs": 2, 00:08:47.503 "num_base_bdevs_discovered": 2, 00:08:47.503 "num_base_bdevs_operational": 2, 00:08:47.503 "base_bdevs_list": [ 00:08:47.503 { 00:08:47.503 "name": "pt1", 00:08:47.503 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:47.503 "is_configured": true, 00:08:47.503 "data_offset": 2048, 00:08:47.503 "data_size": 63488 00:08:47.503 }, 00:08:47.503 { 00:08:47.503 "name": "pt2", 00:08:47.503 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:47.503 "is_configured": true, 00:08:47.503 "data_offset": 2048, 00:08:47.503 "data_size": 63488 00:08:47.503 } 00:08:47.503 ] 00:08:47.503 }' 00:08:47.503 15:14:37 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:47.503 15:14:37 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.761 15:14:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:08:47.761 15:14:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:08:47.761 15:14:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:47.761 15:14:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:47.761 15:14:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:47.761 15:14:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:47.761 15:14:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:47.761 15:14:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:47.761 15:14:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:47.761 15:14:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.761 [2024-11-19 15:14:38.097735] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:48.019 15:14:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:48.019 15:14:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:48.019 "name": "raid_bdev1", 00:08:48.019 "aliases": [ 00:08:48.019 "f5025517-ed2d-46b1-9c3e-d63b904bb272" 00:08:48.019 ], 00:08:48.019 "product_name": "Raid Volume", 00:08:48.019 "block_size": 512, 00:08:48.019 "num_blocks": 126976, 00:08:48.019 "uuid": "f5025517-ed2d-46b1-9c3e-d63b904bb272", 00:08:48.019 "assigned_rate_limits": { 00:08:48.019 "rw_ios_per_sec": 0, 00:08:48.019 "rw_mbytes_per_sec": 0, 00:08:48.019 "r_mbytes_per_sec": 0, 00:08:48.019 "w_mbytes_per_sec": 0 00:08:48.019 }, 00:08:48.019 "claimed": false, 00:08:48.019 "zoned": false, 00:08:48.019 "supported_io_types": { 00:08:48.019 "read": true, 00:08:48.019 "write": true, 00:08:48.019 "unmap": true, 00:08:48.019 "flush": true, 00:08:48.019 "reset": true, 00:08:48.019 "nvme_admin": false, 00:08:48.019 "nvme_io": false, 00:08:48.019 "nvme_io_md": false, 00:08:48.019 "write_zeroes": true, 00:08:48.019 "zcopy": false, 00:08:48.019 "get_zone_info": false, 00:08:48.019 "zone_management": false, 00:08:48.019 "zone_append": false, 00:08:48.019 "compare": false, 00:08:48.019 "compare_and_write": false, 00:08:48.019 "abort": false, 00:08:48.019 "seek_hole": false, 00:08:48.019 "seek_data": false, 00:08:48.019 "copy": false, 00:08:48.019 "nvme_iov_md": false 00:08:48.019 }, 00:08:48.019 "memory_domains": [ 00:08:48.019 { 00:08:48.019 "dma_device_id": "system", 00:08:48.019 "dma_device_type": 1 00:08:48.019 }, 00:08:48.019 { 00:08:48.019 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:48.019 "dma_device_type": 2 00:08:48.019 }, 00:08:48.019 { 00:08:48.019 "dma_device_id": "system", 00:08:48.019 "dma_device_type": 1 00:08:48.019 }, 00:08:48.019 { 00:08:48.019 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:48.019 "dma_device_type": 2 00:08:48.019 } 00:08:48.019 ], 00:08:48.019 "driver_specific": { 00:08:48.019 "raid": { 00:08:48.019 "uuid": "f5025517-ed2d-46b1-9c3e-d63b904bb272", 00:08:48.019 "strip_size_kb": 64, 00:08:48.019 "state": "online", 00:08:48.019 "raid_level": "concat", 00:08:48.019 "superblock": true, 00:08:48.019 "num_base_bdevs": 2, 00:08:48.019 "num_base_bdevs_discovered": 2, 00:08:48.019 "num_base_bdevs_operational": 2, 00:08:48.019 "base_bdevs_list": [ 00:08:48.019 { 00:08:48.019 "name": "pt1", 00:08:48.019 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:48.019 "is_configured": true, 00:08:48.019 "data_offset": 2048, 00:08:48.019 "data_size": 63488 00:08:48.019 }, 00:08:48.019 { 00:08:48.019 "name": "pt2", 00:08:48.019 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:48.019 "is_configured": true, 00:08:48.019 "data_offset": 2048, 00:08:48.019 "data_size": 63488 00:08:48.019 } 00:08:48.019 ] 00:08:48.019 } 00:08:48.019 } 00:08:48.019 }' 00:08:48.019 15:14:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:48.019 15:14:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:08:48.019 pt2' 00:08:48.019 15:14:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:48.019 15:14:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:48.019 15:14:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:48.019 15:14:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:08:48.019 15:14:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:48.019 15:14:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.019 15:14:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:48.019 15:14:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:48.019 15:14:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:48.019 15:14:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:48.019 15:14:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:48.019 15:14:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:48.020 15:14:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:08:48.020 15:14:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:48.020 15:14:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.020 15:14:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:48.020 15:14:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:48.020 15:14:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:48.020 15:14:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:48.020 15:14:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:08:48.020 15:14:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:48.020 15:14:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.020 [2024-11-19 15:14:38.285373] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:48.020 15:14:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:48.020 15:14:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' f5025517-ed2d-46b1-9c3e-d63b904bb272 '!=' f5025517-ed2d-46b1-9c3e-d63b904bb272 ']' 00:08:48.020 15:14:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy concat 00:08:48.020 15:14:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:48.020 15:14:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:48.020 15:14:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 73504 00:08:48.020 15:14:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 73504 ']' 00:08:48.020 15:14:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 73504 00:08:48.020 15:14:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:08:48.020 15:14:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:48.020 15:14:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 73504 00:08:48.278 15:14:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:48.278 15:14:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:48.278 15:14:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 73504' 00:08:48.278 killing process with pid 73504 00:08:48.278 15:14:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 73504 00:08:48.278 [2024-11-19 15:14:38.372014] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:48.278 [2024-11-19 15:14:38.372166] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:48.278 15:14:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 73504 00:08:48.278 [2024-11-19 15:14:38.372253] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:48.278 [2024-11-19 15:14:38.372265] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:08:48.278 [2024-11-19 15:14:38.415244] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:48.536 15:14:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:08:48.536 00:08:48.536 real 0m3.437s 00:08:48.536 user 0m5.179s 00:08:48.536 sys 0m0.769s 00:08:48.536 15:14:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:48.536 ************************************ 00:08:48.536 END TEST raid_superblock_test 00:08:48.536 ************************************ 00:08:48.536 15:14:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.536 15:14:38 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test concat 2 read 00:08:48.536 15:14:38 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:48.536 15:14:38 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:48.536 15:14:38 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:48.536 ************************************ 00:08:48.536 START TEST raid_read_error_test 00:08:48.536 ************************************ 00:08:48.536 15:14:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test concat 2 read 00:08:48.536 15:14:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:08:48.536 15:14:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:08:48.536 15:14:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:08:48.536 15:14:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:08:48.536 15:14:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:48.536 15:14:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:08:48.536 15:14:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:48.536 15:14:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:48.536 15:14:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:08:48.536 15:14:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:48.536 15:14:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:48.536 15:14:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:08:48.536 15:14:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:08:48.536 15:14:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:08:48.536 15:14:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:08:48.536 15:14:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:08:48.536 15:14:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:08:48.536 15:14:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:08:48.536 15:14:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:08:48.536 15:14:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:08:48.536 15:14:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:08:48.536 15:14:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:08:48.536 15:14:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.E9o3efn4EU 00:08:48.536 15:14:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=73704 00:08:48.536 15:14:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 73704 00:08:48.536 15:14:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:08:48.536 15:14:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 73704 ']' 00:08:48.536 15:14:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:48.536 15:14:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:48.536 15:14:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:48.536 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:48.536 15:14:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:48.536 15:14:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.793 [2024-11-19 15:14:38.904080] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:08:48.793 [2024-11-19 15:14:38.904282] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid73704 ] 00:08:48.793 [2024-11-19 15:14:39.057142] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:48.793 [2024-11-19 15:14:39.083080] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:48.793 [2024-11-19 15:14:39.125032] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:48.793 [2024-11-19 15:14:39.125142] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.728 BaseBdev1_malloc 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.728 true 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.728 [2024-11-19 15:14:39.774870] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:08:49.728 [2024-11-19 15:14:39.774923] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:49.728 [2024-11-19 15:14:39.774958] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:08:49.728 [2024-11-19 15:14:39.774966] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:49.728 [2024-11-19 15:14:39.777069] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:49.728 [2024-11-19 15:14:39.777144] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:08:49.728 BaseBdev1 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.728 BaseBdev2_malloc 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.728 true 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.728 [2024-11-19 15:14:39.815443] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:08:49.728 [2024-11-19 15:14:39.815534] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:49.728 [2024-11-19 15:14:39.815557] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:08:49.728 [2024-11-19 15:14:39.815574] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:49.728 [2024-11-19 15:14:39.817746] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:49.728 [2024-11-19 15:14:39.817784] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:08:49.728 BaseBdev2 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.728 [2024-11-19 15:14:39.827487] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:49.728 [2024-11-19 15:14:39.829453] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:49.728 [2024-11-19 15:14:39.829634] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:49.728 [2024-11-19 15:14:39.829647] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:08:49.728 [2024-11-19 15:14:39.829919] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:08:49.728 [2024-11-19 15:14:39.830083] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:49.728 [2024-11-19 15:14:39.830097] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:08:49.728 [2024-11-19 15:14:39.830218] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.728 15:14:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:49.728 "name": "raid_bdev1", 00:08:49.728 "uuid": "8d79f9de-15b8-468d-beae-16f928a45e81", 00:08:49.728 "strip_size_kb": 64, 00:08:49.728 "state": "online", 00:08:49.728 "raid_level": "concat", 00:08:49.728 "superblock": true, 00:08:49.728 "num_base_bdevs": 2, 00:08:49.728 "num_base_bdevs_discovered": 2, 00:08:49.728 "num_base_bdevs_operational": 2, 00:08:49.728 "base_bdevs_list": [ 00:08:49.728 { 00:08:49.728 "name": "BaseBdev1", 00:08:49.728 "uuid": "b1a06eef-2709-53f7-b01e-3725d32c6aa3", 00:08:49.728 "is_configured": true, 00:08:49.728 "data_offset": 2048, 00:08:49.728 "data_size": 63488 00:08:49.728 }, 00:08:49.728 { 00:08:49.728 "name": "BaseBdev2", 00:08:49.728 "uuid": "b0cd7ed3-24ec-5059-8c6e-3cad0082ad30", 00:08:49.728 "is_configured": true, 00:08:49.728 "data_offset": 2048, 00:08:49.728 "data_size": 63488 00:08:49.728 } 00:08:49.728 ] 00:08:49.729 }' 00:08:49.729 15:14:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:49.729 15:14:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.987 15:14:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:08:49.987 15:14:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:08:49.987 [2024-11-19 15:14:40.311046] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:08:50.923 15:14:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:08:50.923 15:14:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.923 15:14:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.923 15:14:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.923 15:14:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:08:50.923 15:14:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:08:50.923 15:14:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:08:50.923 15:14:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:08:50.923 15:14:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:50.923 15:14:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:50.923 15:14:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:50.923 15:14:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:50.923 15:14:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:50.923 15:14:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:50.923 15:14:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:50.923 15:14:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:50.923 15:14:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:50.923 15:14:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:50.923 15:14:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:50.923 15:14:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.923 15:14:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.923 15:14:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.182 15:14:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:51.182 "name": "raid_bdev1", 00:08:51.182 "uuid": "8d79f9de-15b8-468d-beae-16f928a45e81", 00:08:51.182 "strip_size_kb": 64, 00:08:51.182 "state": "online", 00:08:51.182 "raid_level": "concat", 00:08:51.182 "superblock": true, 00:08:51.182 "num_base_bdevs": 2, 00:08:51.182 "num_base_bdevs_discovered": 2, 00:08:51.182 "num_base_bdevs_operational": 2, 00:08:51.182 "base_bdevs_list": [ 00:08:51.182 { 00:08:51.182 "name": "BaseBdev1", 00:08:51.182 "uuid": "b1a06eef-2709-53f7-b01e-3725d32c6aa3", 00:08:51.182 "is_configured": true, 00:08:51.182 "data_offset": 2048, 00:08:51.182 "data_size": 63488 00:08:51.182 }, 00:08:51.182 { 00:08:51.182 "name": "BaseBdev2", 00:08:51.182 "uuid": "b0cd7ed3-24ec-5059-8c6e-3cad0082ad30", 00:08:51.182 "is_configured": true, 00:08:51.182 "data_offset": 2048, 00:08:51.182 "data_size": 63488 00:08:51.182 } 00:08:51.182 ] 00:08:51.182 }' 00:08:51.182 15:14:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:51.182 15:14:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.440 15:14:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:51.440 15:14:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:51.440 15:14:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.440 [2024-11-19 15:14:41.699162] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:51.440 [2024-11-19 15:14:41.699259] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:51.440 [2024-11-19 15:14:41.701871] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:51.441 [2024-11-19 15:14:41.701960] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:51.441 [2024-11-19 15:14:41.702030] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:51.441 [2024-11-19 15:14:41.702075] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:08:51.441 { 00:08:51.441 "results": [ 00:08:51.441 { 00:08:51.441 "job": "raid_bdev1", 00:08:51.441 "core_mask": "0x1", 00:08:51.441 "workload": "randrw", 00:08:51.441 "percentage": 50, 00:08:51.441 "status": "finished", 00:08:51.441 "queue_depth": 1, 00:08:51.441 "io_size": 131072, 00:08:51.441 "runtime": 1.389037, 00:08:51.441 "iops": 16937.63377073469, 00:08:51.441 "mibps": 2117.2042213418363, 00:08:51.441 "io_failed": 1, 00:08:51.441 "io_timeout": 0, 00:08:51.441 "avg_latency_us": 81.70736686122565, 00:08:51.441 "min_latency_us": 26.047161572052403, 00:08:51.441 "max_latency_us": 1445.2262008733624 00:08:51.441 } 00:08:51.441 ], 00:08:51.441 "core_count": 1 00:08:51.441 } 00:08:51.441 15:14:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.441 15:14:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 73704 00:08:51.441 15:14:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 73704 ']' 00:08:51.441 15:14:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 73704 00:08:51.441 15:14:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:08:51.441 15:14:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:51.441 15:14:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 73704 00:08:51.441 15:14:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:51.441 15:14:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:51.441 15:14:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 73704' 00:08:51.441 killing process with pid 73704 00:08:51.441 15:14:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 73704 00:08:51.441 [2024-11-19 15:14:41.747899] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:51.441 15:14:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 73704 00:08:51.441 [2024-11-19 15:14:41.763601] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:51.699 15:14:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.E9o3efn4EU 00:08:51.699 15:14:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:08:51.699 15:14:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:08:51.699 15:14:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.72 00:08:51.699 15:14:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:08:51.699 15:14:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:51.699 15:14:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:51.699 15:14:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.72 != \0\.\0\0 ]] 00:08:51.699 00:08:51.699 real 0m3.167s 00:08:51.699 user 0m4.043s 00:08:51.699 sys 0m0.504s 00:08:51.699 15:14:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:51.699 ************************************ 00:08:51.699 END TEST raid_read_error_test 00:08:51.699 ************************************ 00:08:51.699 15:14:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.699 15:14:42 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test concat 2 write 00:08:51.699 15:14:42 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:51.699 15:14:42 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:51.699 15:14:42 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:51.958 ************************************ 00:08:51.958 START TEST raid_write_error_test 00:08:51.958 ************************************ 00:08:51.958 15:14:42 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test concat 2 write 00:08:51.958 15:14:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:08:51.958 15:14:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:08:51.958 15:14:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:08:51.958 15:14:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:08:51.958 15:14:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:51.958 15:14:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:08:51.958 15:14:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:51.958 15:14:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:51.958 15:14:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:08:51.958 15:14:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:51.958 15:14:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:51.958 15:14:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:08:51.958 15:14:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:08:51.958 15:14:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:08:51.958 15:14:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:08:51.958 15:14:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:08:51.958 15:14:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:08:51.958 15:14:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:08:51.958 15:14:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:08:51.958 15:14:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:08:51.958 15:14:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:08:51.958 15:14:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:08:51.958 15:14:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.lgcyBjBSnA 00:08:51.958 15:14:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=73833 00:08:51.958 15:14:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 73833 00:08:51.958 15:14:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:08:51.958 15:14:42 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 73833 ']' 00:08:51.959 15:14:42 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:51.959 15:14:42 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:51.959 15:14:42 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:51.959 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:51.959 15:14:42 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:51.959 15:14:42 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.959 [2024-11-19 15:14:42.145774] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:08:51.959 [2024-11-19 15:14:42.146000] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid73833 ] 00:08:52.217 [2024-11-19 15:14:42.304937] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:52.217 [2024-11-19 15:14:42.333702] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:52.217 [2024-11-19 15:14:42.376164] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:52.217 [2024-11-19 15:14:42.376199] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.785 BaseBdev1_malloc 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.785 true 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.785 [2024-11-19 15:14:43.046825] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:08:52.785 [2024-11-19 15:14:43.046970] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:52.785 [2024-11-19 15:14:43.047010] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:08:52.785 [2024-11-19 15:14:43.047020] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:52.785 [2024-11-19 15:14:43.049561] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:52.785 [2024-11-19 15:14:43.049600] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:08:52.785 BaseBdev1 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.785 BaseBdev2_malloc 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.785 true 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.785 [2024-11-19 15:14:43.093842] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:08:52.785 [2024-11-19 15:14:43.093949] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:52.785 [2024-11-19 15:14:43.093989] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:08:52.785 [2024-11-19 15:14:43.094026] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:52.785 [2024-11-19 15:14:43.096468] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:52.785 [2024-11-19 15:14:43.096509] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:08:52.785 BaseBdev2 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.785 [2024-11-19 15:14:43.105887] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:52.785 [2024-11-19 15:14:43.108069] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:52.785 [2024-11-19 15:14:43.108329] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:52.785 [2024-11-19 15:14:43.108351] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:08:52.785 [2024-11-19 15:14:43.108637] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:08:52.785 [2024-11-19 15:14:43.108779] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:52.785 [2024-11-19 15:14:43.108792] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:08:52.785 [2024-11-19 15:14:43.108930] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:52.785 15:14:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.044 15:14:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:53.044 15:14:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:53.044 "name": "raid_bdev1", 00:08:53.044 "uuid": "775a205b-d151-4fe1-bf29-055f9bf95cfd", 00:08:53.044 "strip_size_kb": 64, 00:08:53.044 "state": "online", 00:08:53.044 "raid_level": "concat", 00:08:53.044 "superblock": true, 00:08:53.044 "num_base_bdevs": 2, 00:08:53.044 "num_base_bdevs_discovered": 2, 00:08:53.044 "num_base_bdevs_operational": 2, 00:08:53.044 "base_bdevs_list": [ 00:08:53.044 { 00:08:53.044 "name": "BaseBdev1", 00:08:53.044 "uuid": "15736af1-332b-596d-a733-da177bc30af9", 00:08:53.044 "is_configured": true, 00:08:53.044 "data_offset": 2048, 00:08:53.044 "data_size": 63488 00:08:53.044 }, 00:08:53.044 { 00:08:53.044 "name": "BaseBdev2", 00:08:53.044 "uuid": "3ade7471-b0bb-5eed-a2ae-788b177e2c1f", 00:08:53.044 "is_configured": true, 00:08:53.044 "data_offset": 2048, 00:08:53.044 "data_size": 63488 00:08:53.044 } 00:08:53.044 ] 00:08:53.044 }' 00:08:53.044 15:14:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:53.044 15:14:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.302 15:14:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:08:53.302 15:14:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:08:53.561 [2024-11-19 15:14:43.649391] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:08:54.497 15:14:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:08:54.497 15:14:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:54.497 15:14:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.497 15:14:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:54.497 15:14:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:08:54.497 15:14:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:08:54.497 15:14:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:08:54.497 15:14:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:08:54.497 15:14:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:54.497 15:14:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:54.497 15:14:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:54.497 15:14:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:54.497 15:14:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:54.497 15:14:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:54.497 15:14:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:54.497 15:14:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:54.497 15:14:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:54.497 15:14:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:54.497 15:14:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:54.497 15:14:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:54.497 15:14:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.497 15:14:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:54.497 15:14:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:54.497 "name": "raid_bdev1", 00:08:54.497 "uuid": "775a205b-d151-4fe1-bf29-055f9bf95cfd", 00:08:54.497 "strip_size_kb": 64, 00:08:54.497 "state": "online", 00:08:54.497 "raid_level": "concat", 00:08:54.497 "superblock": true, 00:08:54.497 "num_base_bdevs": 2, 00:08:54.497 "num_base_bdevs_discovered": 2, 00:08:54.497 "num_base_bdevs_operational": 2, 00:08:54.497 "base_bdevs_list": [ 00:08:54.497 { 00:08:54.497 "name": "BaseBdev1", 00:08:54.497 "uuid": "15736af1-332b-596d-a733-da177bc30af9", 00:08:54.497 "is_configured": true, 00:08:54.497 "data_offset": 2048, 00:08:54.497 "data_size": 63488 00:08:54.497 }, 00:08:54.497 { 00:08:54.497 "name": "BaseBdev2", 00:08:54.497 "uuid": "3ade7471-b0bb-5eed-a2ae-788b177e2c1f", 00:08:54.497 "is_configured": true, 00:08:54.497 "data_offset": 2048, 00:08:54.497 "data_size": 63488 00:08:54.497 } 00:08:54.497 ] 00:08:54.497 }' 00:08:54.497 15:14:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:54.497 15:14:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.780 15:14:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:54.780 15:14:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:54.780 15:14:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.780 [2024-11-19 15:14:45.025910] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:54.780 [2024-11-19 15:14:45.025956] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:54.780 [2024-11-19 15:14:45.028818] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:54.780 [2024-11-19 15:14:45.028986] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:54.780 [2024-11-19 15:14:45.029063] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:54.780 [2024-11-19 15:14:45.029076] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:08:54.780 { 00:08:54.780 "results": [ 00:08:54.780 { 00:08:54.780 "job": "raid_bdev1", 00:08:54.780 "core_mask": "0x1", 00:08:54.780 "workload": "randrw", 00:08:54.780 "percentage": 50, 00:08:54.780 "status": "finished", 00:08:54.780 "queue_depth": 1, 00:08:54.780 "io_size": 131072, 00:08:54.780 "runtime": 1.377031, 00:08:54.780 "iops": 15368.571949360618, 00:08:54.780 "mibps": 1921.0714936700772, 00:08:54.780 "io_failed": 1, 00:08:54.780 "io_timeout": 0, 00:08:54.780 "avg_latency_us": 90.94961651118857, 00:08:54.780 "min_latency_us": 25.823580786026202, 00:08:54.780 "max_latency_us": 1387.989519650655 00:08:54.780 } 00:08:54.780 ], 00:08:54.780 "core_count": 1 00:08:54.780 } 00:08:54.781 15:14:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:54.781 15:14:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 73833 00:08:54.781 15:14:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 73833 ']' 00:08:54.781 15:14:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 73833 00:08:54.781 15:14:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:08:54.781 15:14:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:54.781 15:14:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 73833 00:08:54.781 killing process with pid 73833 00:08:54.781 15:14:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:54.781 15:14:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:54.781 15:14:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 73833' 00:08:54.781 15:14:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 73833 00:08:54.781 [2024-11-19 15:14:45.060982] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:54.781 15:14:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 73833 00:08:54.781 [2024-11-19 15:14:45.090234] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:55.348 15:14:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.lgcyBjBSnA 00:08:55.348 15:14:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:08:55.348 15:14:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:08:55.348 15:14:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.73 00:08:55.348 15:14:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:08:55.348 15:14:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:55.348 15:14:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:55.348 15:14:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.73 != \0\.\0\0 ]] 00:08:55.348 00:08:55.348 real 0m3.376s 00:08:55.348 user 0m4.275s 00:08:55.348 sys 0m0.540s 00:08:55.348 ************************************ 00:08:55.348 END TEST raid_write_error_test 00:08:55.348 ************************************ 00:08:55.348 15:14:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:55.348 15:14:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.348 15:14:45 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:08:55.348 15:14:45 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid1 2 false 00:08:55.348 15:14:45 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:55.348 15:14:45 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:55.348 15:14:45 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:55.348 ************************************ 00:08:55.348 START TEST raid_state_function_test 00:08:55.348 ************************************ 00:08:55.348 15:14:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 2 false 00:08:55.348 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:08:55.348 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:08:55.348 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:08:55.348 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:55.348 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:55.348 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:55.348 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:55.349 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:55.349 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:55.349 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:55.349 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:55.349 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:55.349 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:08:55.349 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:55.349 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:55.349 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:55.349 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:55.349 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:55.349 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:08:55.349 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:08:55.349 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:08:55.349 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:08:55.349 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=73966 00:08:55.349 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:55.349 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 73966' 00:08:55.349 Process raid pid: 73966 00:08:55.349 15:14:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 73966 00:08:55.349 15:14:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 73966 ']' 00:08:55.349 15:14:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:55.349 15:14:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:55.349 15:14:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:55.349 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:55.349 15:14:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:55.349 15:14:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.349 [2024-11-19 15:14:45.572776] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:08:55.349 [2024-11-19 15:14:45.572989] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:55.609 [2024-11-19 15:14:45.708559] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:55.609 [2024-11-19 15:14:45.750112] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:55.609 [2024-11-19 15:14:45.827073] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:55.609 [2024-11-19 15:14:45.827229] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:56.185 15:14:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:56.185 15:14:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:08:56.185 15:14:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:08:56.185 15:14:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:56.185 15:14:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:56.185 [2024-11-19 15:14:46.438566] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:56.185 [2024-11-19 15:14:46.438647] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:56.185 [2024-11-19 15:14:46.438674] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:56.185 [2024-11-19 15:14:46.438684] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:56.185 15:14:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:56.185 15:14:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:08:56.185 15:14:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:56.185 15:14:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:56.185 15:14:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:56.185 15:14:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:56.185 15:14:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:56.185 15:14:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:56.185 15:14:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:56.185 15:14:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:56.185 15:14:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:56.185 15:14:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:56.185 15:14:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:56.185 15:14:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:56.185 15:14:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:56.185 15:14:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:56.185 15:14:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:56.185 "name": "Existed_Raid", 00:08:56.185 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:56.185 "strip_size_kb": 0, 00:08:56.185 "state": "configuring", 00:08:56.185 "raid_level": "raid1", 00:08:56.185 "superblock": false, 00:08:56.185 "num_base_bdevs": 2, 00:08:56.185 "num_base_bdevs_discovered": 0, 00:08:56.185 "num_base_bdevs_operational": 2, 00:08:56.185 "base_bdevs_list": [ 00:08:56.185 { 00:08:56.185 "name": "BaseBdev1", 00:08:56.185 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:56.185 "is_configured": false, 00:08:56.185 "data_offset": 0, 00:08:56.185 "data_size": 0 00:08:56.185 }, 00:08:56.185 { 00:08:56.185 "name": "BaseBdev2", 00:08:56.185 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:56.185 "is_configured": false, 00:08:56.185 "data_offset": 0, 00:08:56.185 "data_size": 0 00:08:56.185 } 00:08:56.185 ] 00:08:56.185 }' 00:08:56.185 15:14:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:56.185 15:14:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:56.753 15:14:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:56.753 15:14:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:56.753 15:14:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:56.753 [2024-11-19 15:14:46.825877] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:56.753 [2024-11-19 15:14:46.826047] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:08:56.753 15:14:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:56.753 15:14:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:08:56.753 15:14:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:56.753 15:14:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:56.753 [2024-11-19 15:14:46.837851] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:56.753 [2024-11-19 15:14:46.837977] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:56.753 [2024-11-19 15:14:46.838012] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:56.753 [2024-11-19 15:14:46.838053] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:56.753 15:14:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:56.753 15:14:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:56.753 15:14:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:56.753 15:14:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:56.753 [2024-11-19 15:14:46.865279] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:56.753 BaseBdev1 00:08:56.753 15:14:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:56.753 15:14:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:08:56.753 15:14:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:08:56.753 15:14:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:56.753 15:14:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:56.753 15:14:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:56.754 15:14:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:56.754 15:14:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:56.754 15:14:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:56.754 15:14:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:56.754 15:14:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:56.754 15:14:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:56.754 15:14:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:56.754 15:14:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:56.754 [ 00:08:56.754 { 00:08:56.754 "name": "BaseBdev1", 00:08:56.754 "aliases": [ 00:08:56.754 "5d9784d7-af8f-4fec-967a-3e7e5a082591" 00:08:56.754 ], 00:08:56.754 "product_name": "Malloc disk", 00:08:56.754 "block_size": 512, 00:08:56.754 "num_blocks": 65536, 00:08:56.754 "uuid": "5d9784d7-af8f-4fec-967a-3e7e5a082591", 00:08:56.754 "assigned_rate_limits": { 00:08:56.754 "rw_ios_per_sec": 0, 00:08:56.754 "rw_mbytes_per_sec": 0, 00:08:56.754 "r_mbytes_per_sec": 0, 00:08:56.754 "w_mbytes_per_sec": 0 00:08:56.754 }, 00:08:56.754 "claimed": true, 00:08:56.754 "claim_type": "exclusive_write", 00:08:56.754 "zoned": false, 00:08:56.754 "supported_io_types": { 00:08:56.754 "read": true, 00:08:56.754 "write": true, 00:08:56.754 "unmap": true, 00:08:56.754 "flush": true, 00:08:56.754 "reset": true, 00:08:56.754 "nvme_admin": false, 00:08:56.754 "nvme_io": false, 00:08:56.754 "nvme_io_md": false, 00:08:56.754 "write_zeroes": true, 00:08:56.754 "zcopy": true, 00:08:56.754 "get_zone_info": false, 00:08:56.754 "zone_management": false, 00:08:56.754 "zone_append": false, 00:08:56.754 "compare": false, 00:08:56.754 "compare_and_write": false, 00:08:56.754 "abort": true, 00:08:56.754 "seek_hole": false, 00:08:56.754 "seek_data": false, 00:08:56.754 "copy": true, 00:08:56.754 "nvme_iov_md": false 00:08:56.754 }, 00:08:56.754 "memory_domains": [ 00:08:56.754 { 00:08:56.754 "dma_device_id": "system", 00:08:56.754 "dma_device_type": 1 00:08:56.754 }, 00:08:56.754 { 00:08:56.754 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:56.754 "dma_device_type": 2 00:08:56.754 } 00:08:56.754 ], 00:08:56.754 "driver_specific": {} 00:08:56.754 } 00:08:56.754 ] 00:08:56.754 15:14:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:56.754 15:14:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:56.754 15:14:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:08:56.754 15:14:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:56.754 15:14:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:56.754 15:14:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:56.754 15:14:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:56.754 15:14:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:56.754 15:14:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:56.754 15:14:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:56.754 15:14:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:56.754 15:14:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:56.754 15:14:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:56.754 15:14:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:56.754 15:14:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:56.754 15:14:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:56.754 15:14:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:56.754 15:14:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:56.754 "name": "Existed_Raid", 00:08:56.754 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:56.754 "strip_size_kb": 0, 00:08:56.754 "state": "configuring", 00:08:56.754 "raid_level": "raid1", 00:08:56.754 "superblock": false, 00:08:56.754 "num_base_bdevs": 2, 00:08:56.754 "num_base_bdevs_discovered": 1, 00:08:56.754 "num_base_bdevs_operational": 2, 00:08:56.754 "base_bdevs_list": [ 00:08:56.754 { 00:08:56.754 "name": "BaseBdev1", 00:08:56.754 "uuid": "5d9784d7-af8f-4fec-967a-3e7e5a082591", 00:08:56.754 "is_configured": true, 00:08:56.754 "data_offset": 0, 00:08:56.754 "data_size": 65536 00:08:56.754 }, 00:08:56.754 { 00:08:56.754 "name": "BaseBdev2", 00:08:56.754 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:56.754 "is_configured": false, 00:08:56.754 "data_offset": 0, 00:08:56.754 "data_size": 0 00:08:56.754 } 00:08:56.754 ] 00:08:56.754 }' 00:08:56.754 15:14:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:56.754 15:14:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.322 15:14:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:57.322 15:14:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.322 15:14:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.322 [2024-11-19 15:14:47.364620] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:57.322 [2024-11-19 15:14:47.364697] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:08:57.322 15:14:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.322 15:14:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:08:57.322 15:14:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.322 15:14:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.322 [2024-11-19 15:14:47.372596] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:57.322 [2024-11-19 15:14:47.374898] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:57.322 [2024-11-19 15:14:47.374944] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:57.322 15:14:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.322 15:14:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:08:57.322 15:14:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:57.322 15:14:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:08:57.322 15:14:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:57.322 15:14:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:57.322 15:14:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:57.322 15:14:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:57.322 15:14:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:57.322 15:14:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:57.322 15:14:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:57.322 15:14:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:57.322 15:14:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:57.322 15:14:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:57.322 15:14:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:57.322 15:14:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.322 15:14:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.322 15:14:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.322 15:14:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:57.322 "name": "Existed_Raid", 00:08:57.322 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:57.322 "strip_size_kb": 0, 00:08:57.322 "state": "configuring", 00:08:57.322 "raid_level": "raid1", 00:08:57.322 "superblock": false, 00:08:57.322 "num_base_bdevs": 2, 00:08:57.322 "num_base_bdevs_discovered": 1, 00:08:57.322 "num_base_bdevs_operational": 2, 00:08:57.322 "base_bdevs_list": [ 00:08:57.322 { 00:08:57.322 "name": "BaseBdev1", 00:08:57.322 "uuid": "5d9784d7-af8f-4fec-967a-3e7e5a082591", 00:08:57.322 "is_configured": true, 00:08:57.322 "data_offset": 0, 00:08:57.322 "data_size": 65536 00:08:57.322 }, 00:08:57.322 { 00:08:57.322 "name": "BaseBdev2", 00:08:57.322 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:57.322 "is_configured": false, 00:08:57.322 "data_offset": 0, 00:08:57.322 "data_size": 0 00:08:57.322 } 00:08:57.322 ] 00:08:57.322 }' 00:08:57.322 15:14:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:57.322 15:14:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.582 15:14:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:57.582 15:14:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.582 15:14:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.582 [2024-11-19 15:14:47.860947] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:57.582 [2024-11-19 15:14:47.861118] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:57.582 [2024-11-19 15:14:47.861156] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:08:57.582 [2024-11-19 15:14:47.861526] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:08:57.582 [2024-11-19 15:14:47.861742] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:57.582 [2024-11-19 15:14:47.861794] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:08:57.582 [2024-11-19 15:14:47.862134] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:57.582 BaseBdev2 00:08:57.582 15:14:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.582 15:14:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:08:57.582 15:14:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:08:57.582 15:14:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:57.582 15:14:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:57.582 15:14:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:57.582 15:14:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:57.582 15:14:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:57.582 15:14:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.582 15:14:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.582 15:14:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.582 15:14:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:57.582 15:14:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.582 15:14:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.582 [ 00:08:57.582 { 00:08:57.582 "name": "BaseBdev2", 00:08:57.582 "aliases": [ 00:08:57.582 "f39555a8-088c-4b77-bcf0-26ffd748f122" 00:08:57.582 ], 00:08:57.582 "product_name": "Malloc disk", 00:08:57.582 "block_size": 512, 00:08:57.582 "num_blocks": 65536, 00:08:57.582 "uuid": "f39555a8-088c-4b77-bcf0-26ffd748f122", 00:08:57.582 "assigned_rate_limits": { 00:08:57.582 "rw_ios_per_sec": 0, 00:08:57.582 "rw_mbytes_per_sec": 0, 00:08:57.582 "r_mbytes_per_sec": 0, 00:08:57.582 "w_mbytes_per_sec": 0 00:08:57.582 }, 00:08:57.582 "claimed": true, 00:08:57.582 "claim_type": "exclusive_write", 00:08:57.582 "zoned": false, 00:08:57.582 "supported_io_types": { 00:08:57.582 "read": true, 00:08:57.582 "write": true, 00:08:57.582 "unmap": true, 00:08:57.582 "flush": true, 00:08:57.582 "reset": true, 00:08:57.582 "nvme_admin": false, 00:08:57.582 "nvme_io": false, 00:08:57.582 "nvme_io_md": false, 00:08:57.582 "write_zeroes": true, 00:08:57.582 "zcopy": true, 00:08:57.582 "get_zone_info": false, 00:08:57.582 "zone_management": false, 00:08:57.582 "zone_append": false, 00:08:57.582 "compare": false, 00:08:57.582 "compare_and_write": false, 00:08:57.582 "abort": true, 00:08:57.582 "seek_hole": false, 00:08:57.582 "seek_data": false, 00:08:57.582 "copy": true, 00:08:57.582 "nvme_iov_md": false 00:08:57.582 }, 00:08:57.582 "memory_domains": [ 00:08:57.582 { 00:08:57.582 "dma_device_id": "system", 00:08:57.582 "dma_device_type": 1 00:08:57.582 }, 00:08:57.582 { 00:08:57.582 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:57.582 "dma_device_type": 2 00:08:57.582 } 00:08:57.582 ], 00:08:57.582 "driver_specific": {} 00:08:57.582 } 00:08:57.582 ] 00:08:57.582 15:14:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.582 15:14:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:57.582 15:14:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:57.582 15:14:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:57.582 15:14:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:08:57.582 15:14:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:57.582 15:14:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:57.582 15:14:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:57.582 15:14:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:57.582 15:14:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:57.582 15:14:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:57.582 15:14:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:57.582 15:14:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:57.582 15:14:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:57.582 15:14:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:57.582 15:14:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:57.582 15:14:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.582 15:14:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.841 15:14:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.841 15:14:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:57.841 "name": "Existed_Raid", 00:08:57.841 "uuid": "beeffb0c-1bca-4195-8fd5-0e043f70e627", 00:08:57.841 "strip_size_kb": 0, 00:08:57.841 "state": "online", 00:08:57.841 "raid_level": "raid1", 00:08:57.841 "superblock": false, 00:08:57.841 "num_base_bdevs": 2, 00:08:57.841 "num_base_bdevs_discovered": 2, 00:08:57.841 "num_base_bdevs_operational": 2, 00:08:57.841 "base_bdevs_list": [ 00:08:57.841 { 00:08:57.841 "name": "BaseBdev1", 00:08:57.841 "uuid": "5d9784d7-af8f-4fec-967a-3e7e5a082591", 00:08:57.841 "is_configured": true, 00:08:57.841 "data_offset": 0, 00:08:57.841 "data_size": 65536 00:08:57.841 }, 00:08:57.841 { 00:08:57.841 "name": "BaseBdev2", 00:08:57.841 "uuid": "f39555a8-088c-4b77-bcf0-26ffd748f122", 00:08:57.841 "is_configured": true, 00:08:57.841 "data_offset": 0, 00:08:57.841 "data_size": 65536 00:08:57.841 } 00:08:57.841 ] 00:08:57.841 }' 00:08:57.841 15:14:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:57.841 15:14:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.100 15:14:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:08:58.100 15:14:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:58.100 15:14:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:58.100 15:14:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:58.100 15:14:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:58.100 15:14:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:58.100 15:14:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:58.100 15:14:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:58.100 15:14:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.100 15:14:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.100 [2024-11-19 15:14:48.348434] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:58.100 15:14:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.100 15:14:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:58.100 "name": "Existed_Raid", 00:08:58.100 "aliases": [ 00:08:58.100 "beeffb0c-1bca-4195-8fd5-0e043f70e627" 00:08:58.100 ], 00:08:58.100 "product_name": "Raid Volume", 00:08:58.100 "block_size": 512, 00:08:58.100 "num_blocks": 65536, 00:08:58.100 "uuid": "beeffb0c-1bca-4195-8fd5-0e043f70e627", 00:08:58.100 "assigned_rate_limits": { 00:08:58.100 "rw_ios_per_sec": 0, 00:08:58.100 "rw_mbytes_per_sec": 0, 00:08:58.100 "r_mbytes_per_sec": 0, 00:08:58.100 "w_mbytes_per_sec": 0 00:08:58.100 }, 00:08:58.100 "claimed": false, 00:08:58.100 "zoned": false, 00:08:58.100 "supported_io_types": { 00:08:58.100 "read": true, 00:08:58.100 "write": true, 00:08:58.100 "unmap": false, 00:08:58.100 "flush": false, 00:08:58.100 "reset": true, 00:08:58.100 "nvme_admin": false, 00:08:58.100 "nvme_io": false, 00:08:58.100 "nvme_io_md": false, 00:08:58.100 "write_zeroes": true, 00:08:58.100 "zcopy": false, 00:08:58.100 "get_zone_info": false, 00:08:58.100 "zone_management": false, 00:08:58.100 "zone_append": false, 00:08:58.100 "compare": false, 00:08:58.100 "compare_and_write": false, 00:08:58.100 "abort": false, 00:08:58.100 "seek_hole": false, 00:08:58.100 "seek_data": false, 00:08:58.100 "copy": false, 00:08:58.100 "nvme_iov_md": false 00:08:58.100 }, 00:08:58.100 "memory_domains": [ 00:08:58.100 { 00:08:58.100 "dma_device_id": "system", 00:08:58.100 "dma_device_type": 1 00:08:58.100 }, 00:08:58.100 { 00:08:58.100 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:58.100 "dma_device_type": 2 00:08:58.100 }, 00:08:58.100 { 00:08:58.100 "dma_device_id": "system", 00:08:58.100 "dma_device_type": 1 00:08:58.100 }, 00:08:58.100 { 00:08:58.100 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:58.100 "dma_device_type": 2 00:08:58.100 } 00:08:58.100 ], 00:08:58.100 "driver_specific": { 00:08:58.100 "raid": { 00:08:58.100 "uuid": "beeffb0c-1bca-4195-8fd5-0e043f70e627", 00:08:58.100 "strip_size_kb": 0, 00:08:58.100 "state": "online", 00:08:58.100 "raid_level": "raid1", 00:08:58.100 "superblock": false, 00:08:58.100 "num_base_bdevs": 2, 00:08:58.100 "num_base_bdevs_discovered": 2, 00:08:58.100 "num_base_bdevs_operational": 2, 00:08:58.100 "base_bdevs_list": [ 00:08:58.100 { 00:08:58.100 "name": "BaseBdev1", 00:08:58.100 "uuid": "5d9784d7-af8f-4fec-967a-3e7e5a082591", 00:08:58.100 "is_configured": true, 00:08:58.100 "data_offset": 0, 00:08:58.100 "data_size": 65536 00:08:58.100 }, 00:08:58.100 { 00:08:58.100 "name": "BaseBdev2", 00:08:58.100 "uuid": "f39555a8-088c-4b77-bcf0-26ffd748f122", 00:08:58.100 "is_configured": true, 00:08:58.100 "data_offset": 0, 00:08:58.100 "data_size": 65536 00:08:58.100 } 00:08:58.100 ] 00:08:58.100 } 00:08:58.100 } 00:08:58.100 }' 00:08:58.100 15:14:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:58.100 15:14:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:08:58.100 BaseBdev2' 00:08:58.100 15:14:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:58.360 15:14:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:58.360 15:14:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:58.360 15:14:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:08:58.360 15:14:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.360 15:14:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.360 15:14:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:58.360 15:14:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.360 15:14:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:58.360 15:14:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:58.360 15:14:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:58.360 15:14:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:58.360 15:14:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.360 15:14:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.360 15:14:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:58.360 15:14:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.360 15:14:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:58.360 15:14:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:58.360 15:14:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:58.360 15:14:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.360 15:14:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.360 [2024-11-19 15:14:48.567944] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:58.360 15:14:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.360 15:14:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:08:58.360 15:14:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:08:58.360 15:14:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:58.360 15:14:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:08:58.360 15:14:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:08:58.360 15:14:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:08:58.360 15:14:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:58.360 15:14:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:58.360 15:14:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:58.360 15:14:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:58.360 15:14:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:08:58.360 15:14:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:58.360 15:14:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:58.360 15:14:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:58.360 15:14:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:58.360 15:14:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:58.360 15:14:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.360 15:14:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.360 15:14:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:58.360 15:14:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.360 15:14:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:58.360 "name": "Existed_Raid", 00:08:58.360 "uuid": "beeffb0c-1bca-4195-8fd5-0e043f70e627", 00:08:58.360 "strip_size_kb": 0, 00:08:58.360 "state": "online", 00:08:58.360 "raid_level": "raid1", 00:08:58.360 "superblock": false, 00:08:58.360 "num_base_bdevs": 2, 00:08:58.360 "num_base_bdevs_discovered": 1, 00:08:58.360 "num_base_bdevs_operational": 1, 00:08:58.360 "base_bdevs_list": [ 00:08:58.360 { 00:08:58.360 "name": null, 00:08:58.360 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:58.360 "is_configured": false, 00:08:58.360 "data_offset": 0, 00:08:58.360 "data_size": 65536 00:08:58.360 }, 00:08:58.360 { 00:08:58.360 "name": "BaseBdev2", 00:08:58.360 "uuid": "f39555a8-088c-4b77-bcf0-26ffd748f122", 00:08:58.360 "is_configured": true, 00:08:58.360 "data_offset": 0, 00:08:58.360 "data_size": 65536 00:08:58.360 } 00:08:58.360 ] 00:08:58.360 }' 00:08:58.360 15:14:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:58.360 15:14:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.927 15:14:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:08:58.927 15:14:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:58.927 15:14:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:58.927 15:14:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:58.927 15:14:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.927 15:14:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.927 15:14:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.927 15:14:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:58.927 15:14:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:58.927 15:14:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:08:58.927 15:14:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.927 15:14:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.927 [2024-11-19 15:14:49.083892] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:58.927 [2024-11-19 15:14:49.084070] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:58.927 [2024-11-19 15:14:49.105110] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:58.927 [2024-11-19 15:14:49.105239] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:58.927 [2024-11-19 15:14:49.105285] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:08:58.927 15:14:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.927 15:14:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:58.927 15:14:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:58.927 15:14:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:58.927 15:14:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:08:58.927 15:14:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.927 15:14:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.927 15:14:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.927 15:14:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:08:58.927 15:14:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:08:58.927 15:14:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:08:58.927 15:14:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 73966 00:08:58.928 15:14:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 73966 ']' 00:08:58.928 15:14:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 73966 00:08:58.928 15:14:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:08:58.928 15:14:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:58.928 15:14:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 73966 00:08:58.928 15:14:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:58.928 15:14:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:58.928 15:14:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 73966' 00:08:58.928 killing process with pid 73966 00:08:58.928 15:14:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 73966 00:08:58.928 [2024-11-19 15:14:49.189169] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:58.928 15:14:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 73966 00:08:58.928 [2024-11-19 15:14:49.190750] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:59.186 15:14:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:08:59.186 00:08:59.186 real 0m4.023s 00:08:59.186 user 0m6.248s 00:08:59.186 sys 0m0.804s 00:08:59.186 ************************************ 00:08:59.186 END TEST raid_state_function_test 00:08:59.186 ************************************ 00:08:59.186 15:14:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:59.186 15:14:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.444 15:14:49 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid1 2 true 00:08:59.444 15:14:49 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:59.444 15:14:49 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:59.444 15:14:49 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:59.444 ************************************ 00:08:59.444 START TEST raid_state_function_test_sb 00:08:59.444 ************************************ 00:08:59.444 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 2 true 00:08:59.444 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:08:59.444 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:08:59.444 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:08:59.444 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:59.444 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:59.444 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:59.444 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:59.444 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:59.445 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:59.445 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:59.445 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:59.445 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:59.445 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:08:59.445 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:59.445 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:59.445 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:59.445 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:59.445 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:59.445 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:08:59.445 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:08:59.445 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:08:59.445 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:08:59.445 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=74208 00:08:59.445 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:59.445 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 74208' 00:08:59.445 Process raid pid: 74208 00:08:59.445 15:14:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 74208 00:08:59.445 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 74208 ']' 00:08:59.445 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:59.445 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:59.445 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:59.445 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:59.445 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:59.445 15:14:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:59.445 [2024-11-19 15:14:49.666915] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:08:59.445 [2024-11-19 15:14:49.667053] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:59.703 [2024-11-19 15:14:49.823268] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:59.703 [2024-11-19 15:14:49.861278] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:59.703 [2024-11-19 15:14:49.938204] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:59.703 [2024-11-19 15:14:49.938250] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:00.272 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:00.272 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:09:00.272 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:09:00.272 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:00.272 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.272 [2024-11-19 15:14:50.501320] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:00.272 [2024-11-19 15:14:50.501386] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:00.272 [2024-11-19 15:14:50.501396] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:00.272 [2024-11-19 15:14:50.501407] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:00.272 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:00.272 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:09:00.272 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:00.272 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:00.272 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:00.272 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:00.272 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:00.272 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:00.272 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:00.272 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:00.272 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:00.272 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:00.272 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:00.272 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:00.272 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.272 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:00.272 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:00.272 "name": "Existed_Raid", 00:09:00.272 "uuid": "1f98c655-8fe8-4d4e-8ec4-a8718e0f4a36", 00:09:00.272 "strip_size_kb": 0, 00:09:00.272 "state": "configuring", 00:09:00.272 "raid_level": "raid1", 00:09:00.272 "superblock": true, 00:09:00.272 "num_base_bdevs": 2, 00:09:00.272 "num_base_bdevs_discovered": 0, 00:09:00.272 "num_base_bdevs_operational": 2, 00:09:00.272 "base_bdevs_list": [ 00:09:00.272 { 00:09:00.272 "name": "BaseBdev1", 00:09:00.272 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:00.272 "is_configured": false, 00:09:00.272 "data_offset": 0, 00:09:00.272 "data_size": 0 00:09:00.272 }, 00:09:00.272 { 00:09:00.272 "name": "BaseBdev2", 00:09:00.272 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:00.272 "is_configured": false, 00:09:00.272 "data_offset": 0, 00:09:00.272 "data_size": 0 00:09:00.272 } 00:09:00.272 ] 00:09:00.272 }' 00:09:00.272 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:00.272 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.841 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:00.841 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:00.841 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.841 [2024-11-19 15:14:50.884642] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:00.841 [2024-11-19 15:14:50.884716] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:09:00.841 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:00.841 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:09:00.841 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:00.841 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.841 [2024-11-19 15:14:50.896623] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:00.841 [2024-11-19 15:14:50.896715] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:00.841 [2024-11-19 15:14:50.896743] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:00.841 [2024-11-19 15:14:50.896779] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:00.841 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:00.841 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:00.841 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:00.841 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.841 [2024-11-19 15:14:50.923759] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:00.841 BaseBdev1 00:09:00.841 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:00.841 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:09:00.841 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:00.841 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:00.841 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:00.841 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:00.841 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:00.841 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:00.841 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:00.841 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.841 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:00.841 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:00.841 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:00.841 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.841 [ 00:09:00.841 { 00:09:00.841 "name": "BaseBdev1", 00:09:00.841 "aliases": [ 00:09:00.841 "37b817ec-0d0b-43f8-a3ad-1e2108864f9d" 00:09:00.841 ], 00:09:00.841 "product_name": "Malloc disk", 00:09:00.841 "block_size": 512, 00:09:00.841 "num_blocks": 65536, 00:09:00.841 "uuid": "37b817ec-0d0b-43f8-a3ad-1e2108864f9d", 00:09:00.841 "assigned_rate_limits": { 00:09:00.841 "rw_ios_per_sec": 0, 00:09:00.842 "rw_mbytes_per_sec": 0, 00:09:00.842 "r_mbytes_per_sec": 0, 00:09:00.842 "w_mbytes_per_sec": 0 00:09:00.842 }, 00:09:00.842 "claimed": true, 00:09:00.842 "claim_type": "exclusive_write", 00:09:00.842 "zoned": false, 00:09:00.842 "supported_io_types": { 00:09:00.842 "read": true, 00:09:00.842 "write": true, 00:09:00.842 "unmap": true, 00:09:00.842 "flush": true, 00:09:00.842 "reset": true, 00:09:00.842 "nvme_admin": false, 00:09:00.842 "nvme_io": false, 00:09:00.842 "nvme_io_md": false, 00:09:00.842 "write_zeroes": true, 00:09:00.842 "zcopy": true, 00:09:00.842 "get_zone_info": false, 00:09:00.842 "zone_management": false, 00:09:00.842 "zone_append": false, 00:09:00.842 "compare": false, 00:09:00.842 "compare_and_write": false, 00:09:00.842 "abort": true, 00:09:00.842 "seek_hole": false, 00:09:00.842 "seek_data": false, 00:09:00.842 "copy": true, 00:09:00.842 "nvme_iov_md": false 00:09:00.842 }, 00:09:00.842 "memory_domains": [ 00:09:00.842 { 00:09:00.842 "dma_device_id": "system", 00:09:00.842 "dma_device_type": 1 00:09:00.842 }, 00:09:00.842 { 00:09:00.842 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:00.842 "dma_device_type": 2 00:09:00.842 } 00:09:00.842 ], 00:09:00.842 "driver_specific": {} 00:09:00.842 } 00:09:00.842 ] 00:09:00.842 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:00.842 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:00.842 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:09:00.842 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:00.842 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:00.842 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:00.842 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:00.842 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:00.842 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:00.842 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:00.842 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:00.842 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:00.842 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:00.842 15:14:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:00.842 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:00.842 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.842 15:14:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:00.842 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:00.842 "name": "Existed_Raid", 00:09:00.842 "uuid": "010a1d5f-acaa-40ed-b83b-70206c4fe6fb", 00:09:00.842 "strip_size_kb": 0, 00:09:00.842 "state": "configuring", 00:09:00.842 "raid_level": "raid1", 00:09:00.842 "superblock": true, 00:09:00.842 "num_base_bdevs": 2, 00:09:00.842 "num_base_bdevs_discovered": 1, 00:09:00.842 "num_base_bdevs_operational": 2, 00:09:00.842 "base_bdevs_list": [ 00:09:00.842 { 00:09:00.842 "name": "BaseBdev1", 00:09:00.842 "uuid": "37b817ec-0d0b-43f8-a3ad-1e2108864f9d", 00:09:00.842 "is_configured": true, 00:09:00.842 "data_offset": 2048, 00:09:00.842 "data_size": 63488 00:09:00.842 }, 00:09:00.842 { 00:09:00.842 "name": "BaseBdev2", 00:09:00.842 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:00.842 "is_configured": false, 00:09:00.842 "data_offset": 0, 00:09:00.842 "data_size": 0 00:09:00.842 } 00:09:00.842 ] 00:09:00.842 }' 00:09:00.842 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:00.842 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.101 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:01.101 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.101 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.101 [2024-11-19 15:14:51.427026] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:01.101 [2024-11-19 15:14:51.427166] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:09:01.101 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.101 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:09:01.101 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.101 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.101 [2024-11-19 15:14:51.435015] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:01.101 [2024-11-19 15:14:51.437404] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:01.101 [2024-11-19 15:14:51.437485] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:01.360 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.360 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:09:01.360 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:01.360 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:09:01.360 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:01.360 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:01.360 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:01.360 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:01.360 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:01.360 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:01.360 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:01.360 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:01.360 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:01.360 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:01.360 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:01.360 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.360 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.360 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.360 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:01.360 "name": "Existed_Raid", 00:09:01.360 "uuid": "cb1894bf-c576-4633-a301-b193c7f3df64", 00:09:01.360 "strip_size_kb": 0, 00:09:01.360 "state": "configuring", 00:09:01.360 "raid_level": "raid1", 00:09:01.360 "superblock": true, 00:09:01.360 "num_base_bdevs": 2, 00:09:01.360 "num_base_bdevs_discovered": 1, 00:09:01.360 "num_base_bdevs_operational": 2, 00:09:01.360 "base_bdevs_list": [ 00:09:01.360 { 00:09:01.360 "name": "BaseBdev1", 00:09:01.360 "uuid": "37b817ec-0d0b-43f8-a3ad-1e2108864f9d", 00:09:01.360 "is_configured": true, 00:09:01.360 "data_offset": 2048, 00:09:01.360 "data_size": 63488 00:09:01.360 }, 00:09:01.360 { 00:09:01.360 "name": "BaseBdev2", 00:09:01.360 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:01.360 "is_configured": false, 00:09:01.360 "data_offset": 0, 00:09:01.360 "data_size": 0 00:09:01.360 } 00:09:01.360 ] 00:09:01.360 }' 00:09:01.360 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:01.360 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.620 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:01.620 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.620 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.620 [2024-11-19 15:14:51.863112] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:01.620 [2024-11-19 15:14:51.863484] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:09:01.620 [2024-11-19 15:14:51.863555] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:01.620 BaseBdev2 00:09:01.620 [2024-11-19 15:14:51.863924] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:09:01.620 [2024-11-19 15:14:51.864113] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:09:01.620 [2024-11-19 15:14:51.864137] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:09:01.620 [2024-11-19 15:14:51.864272] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:01.620 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.620 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:09:01.620 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:01.620 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:01.620 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:01.620 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:01.620 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:01.620 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:01.620 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.620 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.620 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.620 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:01.620 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.620 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.620 [ 00:09:01.620 { 00:09:01.620 "name": "BaseBdev2", 00:09:01.620 "aliases": [ 00:09:01.620 "c5eb0cec-ffe2-4e52-87b9-001b0788b88c" 00:09:01.620 ], 00:09:01.620 "product_name": "Malloc disk", 00:09:01.620 "block_size": 512, 00:09:01.620 "num_blocks": 65536, 00:09:01.620 "uuid": "c5eb0cec-ffe2-4e52-87b9-001b0788b88c", 00:09:01.620 "assigned_rate_limits": { 00:09:01.620 "rw_ios_per_sec": 0, 00:09:01.620 "rw_mbytes_per_sec": 0, 00:09:01.620 "r_mbytes_per_sec": 0, 00:09:01.620 "w_mbytes_per_sec": 0 00:09:01.620 }, 00:09:01.620 "claimed": true, 00:09:01.620 "claim_type": "exclusive_write", 00:09:01.620 "zoned": false, 00:09:01.620 "supported_io_types": { 00:09:01.620 "read": true, 00:09:01.620 "write": true, 00:09:01.620 "unmap": true, 00:09:01.620 "flush": true, 00:09:01.620 "reset": true, 00:09:01.620 "nvme_admin": false, 00:09:01.620 "nvme_io": false, 00:09:01.620 "nvme_io_md": false, 00:09:01.620 "write_zeroes": true, 00:09:01.620 "zcopy": true, 00:09:01.620 "get_zone_info": false, 00:09:01.620 "zone_management": false, 00:09:01.620 "zone_append": false, 00:09:01.620 "compare": false, 00:09:01.620 "compare_and_write": false, 00:09:01.620 "abort": true, 00:09:01.620 "seek_hole": false, 00:09:01.620 "seek_data": false, 00:09:01.620 "copy": true, 00:09:01.620 "nvme_iov_md": false 00:09:01.620 }, 00:09:01.620 "memory_domains": [ 00:09:01.620 { 00:09:01.620 "dma_device_id": "system", 00:09:01.620 "dma_device_type": 1 00:09:01.620 }, 00:09:01.620 { 00:09:01.620 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:01.620 "dma_device_type": 2 00:09:01.620 } 00:09:01.620 ], 00:09:01.620 "driver_specific": {} 00:09:01.620 } 00:09:01.620 ] 00:09:01.620 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.620 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:01.620 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:01.620 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:01.621 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:09:01.621 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:01.621 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:01.621 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:01.621 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:01.621 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:01.621 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:01.621 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:01.621 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:01.621 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:01.621 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:01.621 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:01.621 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.621 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.621 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.621 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:01.621 "name": "Existed_Raid", 00:09:01.621 "uuid": "cb1894bf-c576-4633-a301-b193c7f3df64", 00:09:01.621 "strip_size_kb": 0, 00:09:01.621 "state": "online", 00:09:01.621 "raid_level": "raid1", 00:09:01.621 "superblock": true, 00:09:01.621 "num_base_bdevs": 2, 00:09:01.621 "num_base_bdevs_discovered": 2, 00:09:01.621 "num_base_bdevs_operational": 2, 00:09:01.621 "base_bdevs_list": [ 00:09:01.621 { 00:09:01.621 "name": "BaseBdev1", 00:09:01.621 "uuid": "37b817ec-0d0b-43f8-a3ad-1e2108864f9d", 00:09:01.621 "is_configured": true, 00:09:01.621 "data_offset": 2048, 00:09:01.621 "data_size": 63488 00:09:01.621 }, 00:09:01.621 { 00:09:01.621 "name": "BaseBdev2", 00:09:01.621 "uuid": "c5eb0cec-ffe2-4e52-87b9-001b0788b88c", 00:09:01.621 "is_configured": true, 00:09:01.621 "data_offset": 2048, 00:09:01.621 "data_size": 63488 00:09:01.621 } 00:09:01.621 ] 00:09:01.621 }' 00:09:01.621 15:14:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:01.621 15:14:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.190 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:09:02.190 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:02.190 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:02.190 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:02.190 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:09:02.190 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:02.190 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:02.190 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:02.190 15:14:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:02.190 15:14:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.190 [2024-11-19 15:14:52.350666] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:02.190 15:14:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:02.190 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:02.190 "name": "Existed_Raid", 00:09:02.190 "aliases": [ 00:09:02.190 "cb1894bf-c576-4633-a301-b193c7f3df64" 00:09:02.190 ], 00:09:02.190 "product_name": "Raid Volume", 00:09:02.190 "block_size": 512, 00:09:02.190 "num_blocks": 63488, 00:09:02.190 "uuid": "cb1894bf-c576-4633-a301-b193c7f3df64", 00:09:02.190 "assigned_rate_limits": { 00:09:02.190 "rw_ios_per_sec": 0, 00:09:02.190 "rw_mbytes_per_sec": 0, 00:09:02.190 "r_mbytes_per_sec": 0, 00:09:02.190 "w_mbytes_per_sec": 0 00:09:02.190 }, 00:09:02.190 "claimed": false, 00:09:02.190 "zoned": false, 00:09:02.190 "supported_io_types": { 00:09:02.190 "read": true, 00:09:02.190 "write": true, 00:09:02.190 "unmap": false, 00:09:02.190 "flush": false, 00:09:02.190 "reset": true, 00:09:02.190 "nvme_admin": false, 00:09:02.190 "nvme_io": false, 00:09:02.190 "nvme_io_md": false, 00:09:02.190 "write_zeroes": true, 00:09:02.190 "zcopy": false, 00:09:02.190 "get_zone_info": false, 00:09:02.191 "zone_management": false, 00:09:02.191 "zone_append": false, 00:09:02.191 "compare": false, 00:09:02.191 "compare_and_write": false, 00:09:02.191 "abort": false, 00:09:02.191 "seek_hole": false, 00:09:02.191 "seek_data": false, 00:09:02.191 "copy": false, 00:09:02.191 "nvme_iov_md": false 00:09:02.191 }, 00:09:02.191 "memory_domains": [ 00:09:02.191 { 00:09:02.191 "dma_device_id": "system", 00:09:02.191 "dma_device_type": 1 00:09:02.191 }, 00:09:02.191 { 00:09:02.191 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:02.191 "dma_device_type": 2 00:09:02.191 }, 00:09:02.191 { 00:09:02.191 "dma_device_id": "system", 00:09:02.191 "dma_device_type": 1 00:09:02.191 }, 00:09:02.191 { 00:09:02.191 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:02.191 "dma_device_type": 2 00:09:02.191 } 00:09:02.191 ], 00:09:02.191 "driver_specific": { 00:09:02.191 "raid": { 00:09:02.191 "uuid": "cb1894bf-c576-4633-a301-b193c7f3df64", 00:09:02.191 "strip_size_kb": 0, 00:09:02.191 "state": "online", 00:09:02.191 "raid_level": "raid1", 00:09:02.191 "superblock": true, 00:09:02.191 "num_base_bdevs": 2, 00:09:02.191 "num_base_bdevs_discovered": 2, 00:09:02.191 "num_base_bdevs_operational": 2, 00:09:02.191 "base_bdevs_list": [ 00:09:02.191 { 00:09:02.191 "name": "BaseBdev1", 00:09:02.191 "uuid": "37b817ec-0d0b-43f8-a3ad-1e2108864f9d", 00:09:02.191 "is_configured": true, 00:09:02.191 "data_offset": 2048, 00:09:02.191 "data_size": 63488 00:09:02.191 }, 00:09:02.191 { 00:09:02.191 "name": "BaseBdev2", 00:09:02.191 "uuid": "c5eb0cec-ffe2-4e52-87b9-001b0788b88c", 00:09:02.191 "is_configured": true, 00:09:02.191 "data_offset": 2048, 00:09:02.191 "data_size": 63488 00:09:02.191 } 00:09:02.191 ] 00:09:02.191 } 00:09:02.191 } 00:09:02.191 }' 00:09:02.191 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:02.191 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:09:02.191 BaseBdev2' 00:09:02.191 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:02.191 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:02.191 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:02.191 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:09:02.191 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:02.191 15:14:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:02.191 15:14:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.191 15:14:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:02.191 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:02.191 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:02.191 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:02.191 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:02.191 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:02.191 15:14:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:02.191 15:14:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.191 15:14:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:02.455 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:02.455 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:02.455 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:02.455 15:14:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:02.455 15:14:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.455 [2024-11-19 15:14:52.534134] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:02.455 15:14:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:02.455 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:09:02.455 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:09:02.455 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:02.455 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:09:02.455 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:09:02.455 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:09:02.455 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:02.455 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:02.455 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:02.455 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:02.455 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:09:02.455 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:02.455 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:02.455 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:02.455 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:02.455 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:02.455 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:02.455 15:14:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:02.455 15:14:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.455 15:14:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:02.455 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:02.455 "name": "Existed_Raid", 00:09:02.455 "uuid": "cb1894bf-c576-4633-a301-b193c7f3df64", 00:09:02.455 "strip_size_kb": 0, 00:09:02.455 "state": "online", 00:09:02.455 "raid_level": "raid1", 00:09:02.455 "superblock": true, 00:09:02.455 "num_base_bdevs": 2, 00:09:02.455 "num_base_bdevs_discovered": 1, 00:09:02.455 "num_base_bdevs_operational": 1, 00:09:02.455 "base_bdevs_list": [ 00:09:02.455 { 00:09:02.455 "name": null, 00:09:02.455 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:02.455 "is_configured": false, 00:09:02.455 "data_offset": 0, 00:09:02.455 "data_size": 63488 00:09:02.455 }, 00:09:02.455 { 00:09:02.455 "name": "BaseBdev2", 00:09:02.455 "uuid": "c5eb0cec-ffe2-4e52-87b9-001b0788b88c", 00:09:02.455 "is_configured": true, 00:09:02.455 "data_offset": 2048, 00:09:02.455 "data_size": 63488 00:09:02.455 } 00:09:02.455 ] 00:09:02.455 }' 00:09:02.455 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:02.455 15:14:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.714 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:09:02.714 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:02.714 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:02.714 15:14:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:02.714 15:14:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:02.714 15:14:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.714 15:14:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:02.714 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:02.714 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:02.714 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:09:02.714 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:02.714 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.714 [2024-11-19 15:14:53.026588] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:02.714 [2024-11-19 15:14:53.026751] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:02.714 [2024-11-19 15:14:53.048069] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:02.714 [2024-11-19 15:14:53.048130] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:02.714 [2024-11-19 15:14:53.048145] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:09:02.714 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:02.714 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:02.714 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:02.973 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:09:02.973 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:02.973 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:02.973 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.973 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:02.973 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:09:02.973 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:09:02.973 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:09:02.973 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 74208 00:09:02.973 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 74208 ']' 00:09:02.973 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 74208 00:09:02.973 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:09:02.973 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:02.973 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 74208 00:09:02.973 killing process with pid 74208 00:09:02.973 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:02.973 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:02.973 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 74208' 00:09:02.973 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 74208 00:09:02.973 [2024-11-19 15:14:53.148250] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:02.973 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 74208 00:09:02.973 [2024-11-19 15:14:53.149903] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:03.231 ************************************ 00:09:03.231 END TEST raid_state_function_test_sb 00:09:03.231 15:14:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:09:03.231 00:09:03.231 real 0m3.897s 00:09:03.231 user 0m5.967s 00:09:03.231 sys 0m0.842s 00:09:03.231 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:03.231 15:14:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:03.231 ************************************ 00:09:03.231 15:14:53 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid1 2 00:09:03.231 15:14:53 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:09:03.231 15:14:53 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:03.231 15:14:53 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:03.231 ************************************ 00:09:03.231 START TEST raid_superblock_test 00:09:03.231 ************************************ 00:09:03.231 15:14:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid1 2 00:09:03.231 15:14:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:09:03.231 15:14:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:09:03.231 15:14:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:09:03.231 15:14:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:09:03.231 15:14:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:09:03.231 15:14:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:09:03.231 15:14:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:09:03.231 15:14:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:09:03.231 15:14:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:09:03.231 15:14:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:09:03.231 15:14:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:09:03.231 15:14:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:09:03.231 15:14:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:09:03.231 15:14:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:09:03.231 15:14:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:09:03.231 15:14:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=74449 00:09:03.231 15:14:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:09:03.231 15:14:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 74449 00:09:03.231 15:14:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 74449 ']' 00:09:03.231 15:14:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:03.231 15:14:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:03.231 15:14:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:03.231 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:03.231 15:14:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:03.231 15:14:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:03.490 [2024-11-19 15:14:53.625593] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:09:03.490 [2024-11-19 15:14:53.625794] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid74449 ] 00:09:03.490 [2024-11-19 15:14:53.778306] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:03.490 [2024-11-19 15:14:53.815628] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:03.748 [2024-11-19 15:14:53.891732] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:03.748 [2024-11-19 15:14:53.891895] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:04.314 15:14:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:04.314 15:14:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:09:04.314 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:09:04.314 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:04.314 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:09:04.314 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:09:04.314 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:09:04.314 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:04.314 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:04.314 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:04.314 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:09:04.314 15:14:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.314 15:14:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.314 malloc1 00:09:04.314 15:14:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.314 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:04.314 15:14:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.314 15:14:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.314 [2024-11-19 15:14:54.493924] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:04.314 [2024-11-19 15:14:54.494057] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:04.314 [2024-11-19 15:14:54.494105] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:09:04.314 [2024-11-19 15:14:54.494162] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:04.314 [2024-11-19 15:14:54.496764] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:04.314 [2024-11-19 15:14:54.496842] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:04.314 pt1 00:09:04.314 15:14:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.314 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:04.314 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:04.314 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:09:04.314 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:09:04.314 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:09:04.314 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:04.314 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:04.314 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:04.314 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:09:04.314 15:14:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.314 15:14:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.314 malloc2 00:09:04.314 15:14:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.314 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:04.314 15:14:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.314 15:14:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.314 [2024-11-19 15:14:54.532602] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:04.314 [2024-11-19 15:14:54.532657] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:04.314 [2024-11-19 15:14:54.532675] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:09:04.314 [2024-11-19 15:14:54.532687] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:04.314 [2024-11-19 15:14:54.535151] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:04.314 [2024-11-19 15:14:54.535189] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:04.314 pt2 00:09:04.314 15:14:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.314 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:04.314 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:04.314 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:09:04.314 15:14:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.314 15:14:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.314 [2024-11-19 15:14:54.544627] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:04.314 [2024-11-19 15:14:54.546964] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:04.314 [2024-11-19 15:14:54.547123] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:09:04.314 [2024-11-19 15:14:54.547146] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:04.314 [2024-11-19 15:14:54.547460] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:09:04.314 [2024-11-19 15:14:54.547650] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:09:04.314 [2024-11-19 15:14:54.547666] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:09:04.314 [2024-11-19 15:14:54.547836] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:04.315 15:14:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.315 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:09:04.315 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:04.315 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:04.315 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:04.315 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:04.315 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:04.315 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:04.315 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:04.315 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:04.315 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:04.315 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:04.315 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:04.315 15:14:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.315 15:14:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.315 15:14:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.315 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:04.315 "name": "raid_bdev1", 00:09:04.315 "uuid": "a0b5d9e9-c609-44da-aaaa-411b79314d62", 00:09:04.315 "strip_size_kb": 0, 00:09:04.315 "state": "online", 00:09:04.315 "raid_level": "raid1", 00:09:04.315 "superblock": true, 00:09:04.315 "num_base_bdevs": 2, 00:09:04.315 "num_base_bdevs_discovered": 2, 00:09:04.315 "num_base_bdevs_operational": 2, 00:09:04.315 "base_bdevs_list": [ 00:09:04.315 { 00:09:04.315 "name": "pt1", 00:09:04.315 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:04.315 "is_configured": true, 00:09:04.315 "data_offset": 2048, 00:09:04.315 "data_size": 63488 00:09:04.315 }, 00:09:04.315 { 00:09:04.315 "name": "pt2", 00:09:04.315 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:04.315 "is_configured": true, 00:09:04.315 "data_offset": 2048, 00:09:04.315 "data_size": 63488 00:09:04.315 } 00:09:04.315 ] 00:09:04.315 }' 00:09:04.315 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:04.315 15:14:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.882 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:09:04.882 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:09:04.882 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:04.882 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:04.882 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:04.882 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:04.882 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:04.882 15:14:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:04.882 15:14:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.882 15:14:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.882 [2024-11-19 15:14:54.980187] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:04.882 15:14:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.882 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:04.882 "name": "raid_bdev1", 00:09:04.882 "aliases": [ 00:09:04.882 "a0b5d9e9-c609-44da-aaaa-411b79314d62" 00:09:04.882 ], 00:09:04.882 "product_name": "Raid Volume", 00:09:04.882 "block_size": 512, 00:09:04.882 "num_blocks": 63488, 00:09:04.882 "uuid": "a0b5d9e9-c609-44da-aaaa-411b79314d62", 00:09:04.882 "assigned_rate_limits": { 00:09:04.882 "rw_ios_per_sec": 0, 00:09:04.882 "rw_mbytes_per_sec": 0, 00:09:04.882 "r_mbytes_per_sec": 0, 00:09:04.882 "w_mbytes_per_sec": 0 00:09:04.882 }, 00:09:04.882 "claimed": false, 00:09:04.882 "zoned": false, 00:09:04.882 "supported_io_types": { 00:09:04.882 "read": true, 00:09:04.882 "write": true, 00:09:04.882 "unmap": false, 00:09:04.882 "flush": false, 00:09:04.882 "reset": true, 00:09:04.882 "nvme_admin": false, 00:09:04.882 "nvme_io": false, 00:09:04.882 "nvme_io_md": false, 00:09:04.882 "write_zeroes": true, 00:09:04.882 "zcopy": false, 00:09:04.882 "get_zone_info": false, 00:09:04.882 "zone_management": false, 00:09:04.882 "zone_append": false, 00:09:04.882 "compare": false, 00:09:04.882 "compare_and_write": false, 00:09:04.882 "abort": false, 00:09:04.882 "seek_hole": false, 00:09:04.882 "seek_data": false, 00:09:04.882 "copy": false, 00:09:04.882 "nvme_iov_md": false 00:09:04.882 }, 00:09:04.882 "memory_domains": [ 00:09:04.882 { 00:09:04.882 "dma_device_id": "system", 00:09:04.882 "dma_device_type": 1 00:09:04.882 }, 00:09:04.882 { 00:09:04.882 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:04.882 "dma_device_type": 2 00:09:04.882 }, 00:09:04.882 { 00:09:04.882 "dma_device_id": "system", 00:09:04.882 "dma_device_type": 1 00:09:04.882 }, 00:09:04.882 { 00:09:04.882 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:04.882 "dma_device_type": 2 00:09:04.882 } 00:09:04.882 ], 00:09:04.882 "driver_specific": { 00:09:04.882 "raid": { 00:09:04.882 "uuid": "a0b5d9e9-c609-44da-aaaa-411b79314d62", 00:09:04.882 "strip_size_kb": 0, 00:09:04.882 "state": "online", 00:09:04.882 "raid_level": "raid1", 00:09:04.882 "superblock": true, 00:09:04.882 "num_base_bdevs": 2, 00:09:04.882 "num_base_bdevs_discovered": 2, 00:09:04.882 "num_base_bdevs_operational": 2, 00:09:04.882 "base_bdevs_list": [ 00:09:04.882 { 00:09:04.882 "name": "pt1", 00:09:04.882 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:04.882 "is_configured": true, 00:09:04.882 "data_offset": 2048, 00:09:04.882 "data_size": 63488 00:09:04.882 }, 00:09:04.882 { 00:09:04.882 "name": "pt2", 00:09:04.882 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:04.882 "is_configured": true, 00:09:04.882 "data_offset": 2048, 00:09:04.882 "data_size": 63488 00:09:04.882 } 00:09:04.882 ] 00:09:04.882 } 00:09:04.882 } 00:09:04.882 }' 00:09:04.882 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:04.882 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:09:04.882 pt2' 00:09:04.882 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:04.882 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:04.882 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:04.882 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:09:04.882 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:04.882 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.882 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.882 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.882 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:04.882 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:04.882 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:04.882 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:09:04.882 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:04.882 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.882 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.882 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.141 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:05.141 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:05.141 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:05.141 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:09:05.141 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.141 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.141 [2024-11-19 15:14:55.235696] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:05.141 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.141 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=a0b5d9e9-c609-44da-aaaa-411b79314d62 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z a0b5d9e9-c609-44da-aaaa-411b79314d62 ']' 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.142 [2024-11-19 15:14:55.283410] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:05.142 [2024-11-19 15:14:55.283437] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:05.142 [2024-11-19 15:14:55.283512] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:05.142 [2024-11-19 15:14:55.283579] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:05.142 [2024-11-19 15:14:55.283595] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.142 [2024-11-19 15:14:55.403231] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:09:05.142 [2024-11-19 15:14:55.405429] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:09:05.142 [2024-11-19 15:14:55.405579] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:09:05.142 [2024-11-19 15:14:55.405638] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:09:05.142 [2024-11-19 15:14:55.405658] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:05.142 [2024-11-19 15:14:55.405673] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:09:05.142 request: 00:09:05.142 { 00:09:05.142 "name": "raid_bdev1", 00:09:05.142 "raid_level": "raid1", 00:09:05.142 "base_bdevs": [ 00:09:05.142 "malloc1", 00:09:05.142 "malloc2" 00:09:05.142 ], 00:09:05.142 "superblock": false, 00:09:05.142 "method": "bdev_raid_create", 00:09:05.142 "req_id": 1 00:09:05.142 } 00:09:05.142 Got JSON-RPC error response 00:09:05.142 response: 00:09:05.142 { 00:09:05.142 "code": -17, 00:09:05.142 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:09:05.142 } 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.142 [2024-11-19 15:14:55.459113] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:05.142 [2024-11-19 15:14:55.459201] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:05.142 [2024-11-19 15:14:55.459238] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:09:05.142 [2024-11-19 15:14:55.459265] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:05.142 [2024-11-19 15:14:55.461754] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:05.142 [2024-11-19 15:14:55.461840] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:05.142 [2024-11-19 15:14:55.461928] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:09:05.142 [2024-11-19 15:14:55.461997] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:05.142 pt1 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.142 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.401 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.401 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:05.401 "name": "raid_bdev1", 00:09:05.401 "uuid": "a0b5d9e9-c609-44da-aaaa-411b79314d62", 00:09:05.401 "strip_size_kb": 0, 00:09:05.401 "state": "configuring", 00:09:05.401 "raid_level": "raid1", 00:09:05.401 "superblock": true, 00:09:05.401 "num_base_bdevs": 2, 00:09:05.401 "num_base_bdevs_discovered": 1, 00:09:05.401 "num_base_bdevs_operational": 2, 00:09:05.401 "base_bdevs_list": [ 00:09:05.401 { 00:09:05.401 "name": "pt1", 00:09:05.401 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:05.401 "is_configured": true, 00:09:05.401 "data_offset": 2048, 00:09:05.401 "data_size": 63488 00:09:05.401 }, 00:09:05.401 { 00:09:05.401 "name": null, 00:09:05.401 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:05.401 "is_configured": false, 00:09:05.401 "data_offset": 2048, 00:09:05.401 "data_size": 63488 00:09:05.401 } 00:09:05.401 ] 00:09:05.401 }' 00:09:05.401 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:05.401 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.660 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:09:05.660 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:09:05.660 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:05.660 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:05.660 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.660 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.660 [2024-11-19 15:14:55.906330] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:05.660 [2024-11-19 15:14:55.906375] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:05.660 [2024-11-19 15:14:55.906395] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:09:05.660 [2024-11-19 15:14:55.906403] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:05.660 [2024-11-19 15:14:55.906750] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:05.660 [2024-11-19 15:14:55.906766] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:05.660 [2024-11-19 15:14:55.906823] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:05.660 [2024-11-19 15:14:55.906840] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:05.660 [2024-11-19 15:14:55.906933] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:09:05.660 [2024-11-19 15:14:55.906943] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:05.660 [2024-11-19 15:14:55.907226] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:09:05.660 [2024-11-19 15:14:55.907346] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:09:05.660 [2024-11-19 15:14:55.907362] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:09:05.660 [2024-11-19 15:14:55.907460] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:05.660 pt2 00:09:05.660 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.660 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:09:05.660 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:05.660 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:09:05.660 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:05.660 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:05.660 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:05.660 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:05.660 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:05.660 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:05.660 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:05.660 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:05.660 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:05.660 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:05.660 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:05.660 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.660 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.660 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.660 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:05.660 "name": "raid_bdev1", 00:09:05.660 "uuid": "a0b5d9e9-c609-44da-aaaa-411b79314d62", 00:09:05.660 "strip_size_kb": 0, 00:09:05.660 "state": "online", 00:09:05.660 "raid_level": "raid1", 00:09:05.660 "superblock": true, 00:09:05.660 "num_base_bdevs": 2, 00:09:05.660 "num_base_bdevs_discovered": 2, 00:09:05.660 "num_base_bdevs_operational": 2, 00:09:05.660 "base_bdevs_list": [ 00:09:05.660 { 00:09:05.660 "name": "pt1", 00:09:05.660 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:05.660 "is_configured": true, 00:09:05.660 "data_offset": 2048, 00:09:05.660 "data_size": 63488 00:09:05.660 }, 00:09:05.660 { 00:09:05.660 "name": "pt2", 00:09:05.660 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:05.660 "is_configured": true, 00:09:05.660 "data_offset": 2048, 00:09:05.660 "data_size": 63488 00:09:05.660 } 00:09:05.660 ] 00:09:05.660 }' 00:09:05.660 15:14:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:05.660 15:14:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:06.228 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:09:06.228 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:09:06.228 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:06.228 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:06.228 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:06.228 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:06.228 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:06.228 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:06.228 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:06.228 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:06.228 [2024-11-19 15:14:56.326259] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:06.228 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:06.228 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:06.228 "name": "raid_bdev1", 00:09:06.228 "aliases": [ 00:09:06.228 "a0b5d9e9-c609-44da-aaaa-411b79314d62" 00:09:06.228 ], 00:09:06.228 "product_name": "Raid Volume", 00:09:06.228 "block_size": 512, 00:09:06.228 "num_blocks": 63488, 00:09:06.228 "uuid": "a0b5d9e9-c609-44da-aaaa-411b79314d62", 00:09:06.228 "assigned_rate_limits": { 00:09:06.228 "rw_ios_per_sec": 0, 00:09:06.228 "rw_mbytes_per_sec": 0, 00:09:06.228 "r_mbytes_per_sec": 0, 00:09:06.228 "w_mbytes_per_sec": 0 00:09:06.228 }, 00:09:06.228 "claimed": false, 00:09:06.228 "zoned": false, 00:09:06.228 "supported_io_types": { 00:09:06.228 "read": true, 00:09:06.228 "write": true, 00:09:06.228 "unmap": false, 00:09:06.228 "flush": false, 00:09:06.228 "reset": true, 00:09:06.228 "nvme_admin": false, 00:09:06.228 "nvme_io": false, 00:09:06.228 "nvme_io_md": false, 00:09:06.228 "write_zeroes": true, 00:09:06.228 "zcopy": false, 00:09:06.228 "get_zone_info": false, 00:09:06.228 "zone_management": false, 00:09:06.228 "zone_append": false, 00:09:06.228 "compare": false, 00:09:06.228 "compare_and_write": false, 00:09:06.228 "abort": false, 00:09:06.228 "seek_hole": false, 00:09:06.228 "seek_data": false, 00:09:06.228 "copy": false, 00:09:06.228 "nvme_iov_md": false 00:09:06.228 }, 00:09:06.228 "memory_domains": [ 00:09:06.228 { 00:09:06.228 "dma_device_id": "system", 00:09:06.228 "dma_device_type": 1 00:09:06.228 }, 00:09:06.228 { 00:09:06.228 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:06.228 "dma_device_type": 2 00:09:06.228 }, 00:09:06.228 { 00:09:06.228 "dma_device_id": "system", 00:09:06.228 "dma_device_type": 1 00:09:06.228 }, 00:09:06.228 { 00:09:06.228 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:06.228 "dma_device_type": 2 00:09:06.228 } 00:09:06.228 ], 00:09:06.228 "driver_specific": { 00:09:06.228 "raid": { 00:09:06.228 "uuid": "a0b5d9e9-c609-44da-aaaa-411b79314d62", 00:09:06.228 "strip_size_kb": 0, 00:09:06.228 "state": "online", 00:09:06.228 "raid_level": "raid1", 00:09:06.228 "superblock": true, 00:09:06.228 "num_base_bdevs": 2, 00:09:06.228 "num_base_bdevs_discovered": 2, 00:09:06.228 "num_base_bdevs_operational": 2, 00:09:06.228 "base_bdevs_list": [ 00:09:06.228 { 00:09:06.228 "name": "pt1", 00:09:06.228 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:06.228 "is_configured": true, 00:09:06.228 "data_offset": 2048, 00:09:06.228 "data_size": 63488 00:09:06.228 }, 00:09:06.228 { 00:09:06.228 "name": "pt2", 00:09:06.228 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:06.229 "is_configured": true, 00:09:06.229 "data_offset": 2048, 00:09:06.229 "data_size": 63488 00:09:06.229 } 00:09:06.229 ] 00:09:06.229 } 00:09:06.229 } 00:09:06.229 }' 00:09:06.229 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:06.229 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:09:06.229 pt2' 00:09:06.229 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:06.229 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:06.229 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:06.229 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:09:06.229 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:06.229 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:06.229 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:06.229 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:06.229 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:06.229 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:06.229 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:06.229 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:09:06.229 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:06.229 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:06.229 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:06.229 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:06.229 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:06.229 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:06.229 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:06.229 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:06.229 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:09:06.229 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:06.229 [2024-11-19 15:14:56.561757] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:06.492 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:06.492 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' a0b5d9e9-c609-44da-aaaa-411b79314d62 '!=' a0b5d9e9-c609-44da-aaaa-411b79314d62 ']' 00:09:06.492 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:09:06.492 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:06.492 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:09:06.492 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:09:06.492 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:06.492 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:06.492 [2024-11-19 15:14:56.605453] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:09:06.492 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:06.492 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:09:06.492 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:06.492 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:06.492 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:06.492 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:06.492 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:09:06.492 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:06.492 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:06.492 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:06.492 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:06.492 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:06.492 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:06.492 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:06.492 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:06.492 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:06.492 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:06.492 "name": "raid_bdev1", 00:09:06.492 "uuid": "a0b5d9e9-c609-44da-aaaa-411b79314d62", 00:09:06.492 "strip_size_kb": 0, 00:09:06.492 "state": "online", 00:09:06.492 "raid_level": "raid1", 00:09:06.492 "superblock": true, 00:09:06.492 "num_base_bdevs": 2, 00:09:06.492 "num_base_bdevs_discovered": 1, 00:09:06.492 "num_base_bdevs_operational": 1, 00:09:06.492 "base_bdevs_list": [ 00:09:06.492 { 00:09:06.492 "name": null, 00:09:06.493 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:06.493 "is_configured": false, 00:09:06.493 "data_offset": 0, 00:09:06.493 "data_size": 63488 00:09:06.493 }, 00:09:06.493 { 00:09:06.493 "name": "pt2", 00:09:06.493 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:06.493 "is_configured": true, 00:09:06.493 "data_offset": 2048, 00:09:06.493 "data_size": 63488 00:09:06.493 } 00:09:06.493 ] 00:09:06.493 }' 00:09:06.493 15:14:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:06.493 15:14:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:06.751 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:06.751 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:06.751 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:06.751 [2024-11-19 15:14:57.072703] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:06.751 [2024-11-19 15:14:57.072756] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:06.751 [2024-11-19 15:14:57.072875] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:06.751 [2024-11-19 15:14:57.072936] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:06.751 [2024-11-19 15:14:57.072947] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:09:06.751 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:06.751 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:06.751 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:06.751 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:09:06.751 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.011 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:07.011 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:09:07.011 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:09:07.011 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:09:07.011 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:09:07.011 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:09:07.011 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:07.011 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.011 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:07.011 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:09:07.011 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:09:07.011 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:09:07.011 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:09:07.011 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@519 -- # i=1 00:09:07.011 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:07.011 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:07.011 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.011 [2024-11-19 15:14:57.148509] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:07.011 [2024-11-19 15:14:57.148648] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:07.011 [2024-11-19 15:14:57.148677] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:09:07.011 [2024-11-19 15:14:57.148688] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:07.011 [2024-11-19 15:14:57.151285] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:07.011 [2024-11-19 15:14:57.151321] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:07.011 [2024-11-19 15:14:57.151408] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:07.012 [2024-11-19 15:14:57.151446] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:07.012 [2024-11-19 15:14:57.151534] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:09:07.012 [2024-11-19 15:14:57.151542] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:07.012 [2024-11-19 15:14:57.151812] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:09:07.012 [2024-11-19 15:14:57.151994] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:09:07.012 [2024-11-19 15:14:57.152010] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:09:07.012 [2024-11-19 15:14:57.152136] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:07.012 pt2 00:09:07.012 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:07.012 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:09:07.012 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:07.012 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:07.012 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:07.012 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:07.012 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:09:07.012 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:07.012 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:07.012 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:07.012 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:07.012 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:07.012 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:07.012 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:07.012 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.012 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:07.012 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:07.012 "name": "raid_bdev1", 00:09:07.012 "uuid": "a0b5d9e9-c609-44da-aaaa-411b79314d62", 00:09:07.012 "strip_size_kb": 0, 00:09:07.012 "state": "online", 00:09:07.012 "raid_level": "raid1", 00:09:07.012 "superblock": true, 00:09:07.012 "num_base_bdevs": 2, 00:09:07.012 "num_base_bdevs_discovered": 1, 00:09:07.012 "num_base_bdevs_operational": 1, 00:09:07.012 "base_bdevs_list": [ 00:09:07.012 { 00:09:07.012 "name": null, 00:09:07.012 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:07.012 "is_configured": false, 00:09:07.012 "data_offset": 2048, 00:09:07.012 "data_size": 63488 00:09:07.012 }, 00:09:07.012 { 00:09:07.012 "name": "pt2", 00:09:07.012 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:07.012 "is_configured": true, 00:09:07.012 "data_offset": 2048, 00:09:07.012 "data_size": 63488 00:09:07.012 } 00:09:07.012 ] 00:09:07.012 }' 00:09:07.012 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:07.012 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.582 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:07.582 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:07.582 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.582 [2024-11-19 15:14:57.623919] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:07.582 [2024-11-19 15:14:57.624024] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:07.582 [2024-11-19 15:14:57.624126] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:07.582 [2024-11-19 15:14:57.624202] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:07.582 [2024-11-19 15:14:57.624215] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:09:07.582 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:07.582 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:07.582 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:07.582 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:09:07.582 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.582 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:07.582 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:09:07.582 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:09:07.582 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:09:07.582 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:07.582 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:07.582 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.582 [2024-11-19 15:14:57.663826] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:07.582 [2024-11-19 15:14:57.663905] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:07.582 [2024-11-19 15:14:57.663927] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008d80 00:09:07.583 [2024-11-19 15:14:57.663942] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:07.583 [2024-11-19 15:14:57.666550] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:07.583 [2024-11-19 15:14:57.666634] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:07.583 [2024-11-19 15:14:57.666747] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:09:07.583 [2024-11-19 15:14:57.666800] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:07.583 [2024-11-19 15:14:57.666932] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:09:07.583 [2024-11-19 15:14:57.666951] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:07.583 [2024-11-19 15:14:57.666971] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:09:07.583 [2024-11-19 15:14:57.667018] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:07.583 [2024-11-19 15:14:57.667098] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:09:07.583 [2024-11-19 15:14:57.667110] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:07.583 [2024-11-19 15:14:57.667363] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:09:07.583 [2024-11-19 15:14:57.667497] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:09:07.583 [2024-11-19 15:14:57.667506] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:09:07.583 [2024-11-19 15:14:57.667629] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:07.583 pt1 00:09:07.583 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:07.583 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:09:07.583 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:09:07.583 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:07.583 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:07.583 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:07.583 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:07.583 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:09:07.583 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:07.583 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:07.583 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:07.583 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:07.583 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:07.583 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:07.583 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:07.583 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.583 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:07.583 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:07.583 "name": "raid_bdev1", 00:09:07.583 "uuid": "a0b5d9e9-c609-44da-aaaa-411b79314d62", 00:09:07.583 "strip_size_kb": 0, 00:09:07.583 "state": "online", 00:09:07.583 "raid_level": "raid1", 00:09:07.583 "superblock": true, 00:09:07.583 "num_base_bdevs": 2, 00:09:07.583 "num_base_bdevs_discovered": 1, 00:09:07.583 "num_base_bdevs_operational": 1, 00:09:07.583 "base_bdevs_list": [ 00:09:07.583 { 00:09:07.583 "name": null, 00:09:07.583 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:07.583 "is_configured": false, 00:09:07.583 "data_offset": 2048, 00:09:07.583 "data_size": 63488 00:09:07.583 }, 00:09:07.583 { 00:09:07.583 "name": "pt2", 00:09:07.583 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:07.583 "is_configured": true, 00:09:07.583 "data_offset": 2048, 00:09:07.583 "data_size": 63488 00:09:07.583 } 00:09:07.583 ] 00:09:07.583 }' 00:09:07.583 15:14:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:07.583 15:14:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.842 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:09:07.842 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:07.842 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.842 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:09:07.842 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:07.842 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:09:07.842 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:09:07.842 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:07.842 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:07.842 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.842 [2024-11-19 15:14:58.147382] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:07.842 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:08.102 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' a0b5d9e9-c609-44da-aaaa-411b79314d62 '!=' a0b5d9e9-c609-44da-aaaa-411b79314d62 ']' 00:09:08.102 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 74449 00:09:08.102 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 74449 ']' 00:09:08.102 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 74449 00:09:08.102 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:09:08.102 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:08.102 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 74449 00:09:08.102 killing process with pid 74449 00:09:08.102 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:08.102 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:08.102 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 74449' 00:09:08.102 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 74449 00:09:08.102 [2024-11-19 15:14:58.225585] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:08.102 [2024-11-19 15:14:58.225806] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:08.102 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 74449 00:09:08.102 [2024-11-19 15:14:58.225897] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:08.102 [2024-11-19 15:14:58.225948] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:09:08.102 [2024-11-19 15:14:58.267761] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:08.362 15:14:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:09:08.362 00:09:08.362 real 0m5.055s 00:09:08.362 user 0m8.150s 00:09:08.362 sys 0m1.083s 00:09:08.362 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:08.362 ************************************ 00:09:08.362 END TEST raid_superblock_test 00:09:08.362 ************************************ 00:09:08.362 15:14:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.362 15:14:58 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid1 2 read 00:09:08.362 15:14:58 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:08.362 15:14:58 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:08.362 15:14:58 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:08.362 ************************************ 00:09:08.362 START TEST raid_read_error_test 00:09:08.362 ************************************ 00:09:08.362 15:14:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid1 2 read 00:09:08.362 15:14:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:09:08.362 15:14:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:09:08.362 15:14:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:09:08.362 15:14:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:09:08.362 15:14:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:08.362 15:14:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:09:08.362 15:14:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:08.362 15:14:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:08.362 15:14:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:09:08.362 15:14:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:08.362 15:14:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:08.362 15:14:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:09:08.362 15:14:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:09:08.362 15:14:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:09:08.362 15:14:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:09:08.362 15:14:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:09:08.362 15:14:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:09:08.362 15:14:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:09:08.362 15:14:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:09:08.362 15:14:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:09:08.362 15:14:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:09:08.362 15:14:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.x1oRtBTOgW 00:09:08.362 15:14:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=74763 00:09:08.362 15:14:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:09:08.362 15:14:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 74763 00:09:08.362 15:14:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 74763 ']' 00:09:08.362 15:14:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:08.362 15:14:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:08.362 15:14:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:08.362 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:08.362 15:14:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:08.362 15:14:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.623 [2024-11-19 15:14:58.765143] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:09:08.623 [2024-11-19 15:14:58.765332] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid74763 ] 00:09:08.623 [2024-11-19 15:14:58.902910] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:08.623 [2024-11-19 15:14:58.941080] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:08.882 [2024-11-19 15:14:59.017284] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:08.882 [2024-11-19 15:14:59.017420] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:09.450 15:14:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:09.450 15:14:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:09:09.450 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:09.450 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:09:09.450 15:14:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:09.450 15:14:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.450 BaseBdev1_malloc 00:09:09.450 15:14:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:09.450 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:09:09.450 15:14:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:09.450 15:14:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.450 true 00:09:09.450 15:14:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:09.450 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:09:09.450 15:14:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:09.450 15:14:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.450 [2024-11-19 15:14:59.640054] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:09:09.450 [2024-11-19 15:14:59.640210] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:09.450 [2024-11-19 15:14:59.640245] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:09:09.450 [2024-11-19 15:14:59.640265] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:09.450 [2024-11-19 15:14:59.643055] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:09.450 [2024-11-19 15:14:59.643091] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:09:09.450 BaseBdev1 00:09:09.450 15:14:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:09.450 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:09.450 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:09:09.450 15:14:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:09.450 15:14:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.450 BaseBdev2_malloc 00:09:09.450 15:14:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:09.450 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:09:09.450 15:14:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:09.450 15:14:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.450 true 00:09:09.450 15:14:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:09.450 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:09:09.450 15:14:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:09.450 15:14:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.450 [2024-11-19 15:14:59.687040] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:09:09.450 [2024-11-19 15:14:59.687113] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:09.450 [2024-11-19 15:14:59.687137] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:09:09.450 [2024-11-19 15:14:59.687157] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:09.450 [2024-11-19 15:14:59.689796] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:09.450 [2024-11-19 15:14:59.689840] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:09:09.450 BaseBdev2 00:09:09.450 15:14:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:09.450 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:09:09.450 15:14:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:09.450 15:14:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.450 [2024-11-19 15:14:59.699047] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:09.450 [2024-11-19 15:14:59.701285] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:09.450 [2024-11-19 15:14:59.701556] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:09:09.450 [2024-11-19 15:14:59.701573] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:09.450 [2024-11-19 15:14:59.701881] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:09:09.450 [2024-11-19 15:14:59.702070] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:09:09.450 [2024-11-19 15:14:59.702084] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:09:09.450 [2024-11-19 15:14:59.702227] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:09.450 15:14:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:09.450 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:09:09.450 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:09.450 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:09.450 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:09.450 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:09.450 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:09.451 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:09.451 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:09.451 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:09.451 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:09.451 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:09.451 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:09.451 15:14:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:09.451 15:14:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.451 15:14:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:09.451 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:09.451 "name": "raid_bdev1", 00:09:09.451 "uuid": "15aec222-8164-4f7f-8db1-ed65d33f9887", 00:09:09.451 "strip_size_kb": 0, 00:09:09.451 "state": "online", 00:09:09.451 "raid_level": "raid1", 00:09:09.451 "superblock": true, 00:09:09.451 "num_base_bdevs": 2, 00:09:09.451 "num_base_bdevs_discovered": 2, 00:09:09.451 "num_base_bdevs_operational": 2, 00:09:09.451 "base_bdevs_list": [ 00:09:09.451 { 00:09:09.451 "name": "BaseBdev1", 00:09:09.451 "uuid": "cc90be56-ed87-572c-86ab-b3763dc22085", 00:09:09.451 "is_configured": true, 00:09:09.451 "data_offset": 2048, 00:09:09.451 "data_size": 63488 00:09:09.451 }, 00:09:09.451 { 00:09:09.451 "name": "BaseBdev2", 00:09:09.451 "uuid": "82b4866a-35c4-54de-801f-75f2bfc055d2", 00:09:09.451 "is_configured": true, 00:09:09.451 "data_offset": 2048, 00:09:09.451 "data_size": 63488 00:09:09.451 } 00:09:09.451 ] 00:09:09.451 }' 00:09:09.451 15:14:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:09.451 15:14:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.034 15:15:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:09:10.034 15:15:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:09:10.035 [2024-11-19 15:15:00.258594] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:09:10.975 15:15:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:09:10.975 15:15:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:10.975 15:15:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.975 15:15:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:10.975 15:15:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:09:10.975 15:15:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:09:10.975 15:15:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ read = \w\r\i\t\e ]] 00:09:10.975 15:15:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:09:10.975 15:15:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:09:10.975 15:15:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:10.975 15:15:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:10.975 15:15:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:10.975 15:15:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:10.975 15:15:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:10.975 15:15:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:10.975 15:15:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:10.975 15:15:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:10.975 15:15:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:10.975 15:15:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:10.975 15:15:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:10.975 15:15:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:10.975 15:15:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.975 15:15:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:10.975 15:15:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:10.975 "name": "raid_bdev1", 00:09:10.975 "uuid": "15aec222-8164-4f7f-8db1-ed65d33f9887", 00:09:10.975 "strip_size_kb": 0, 00:09:10.975 "state": "online", 00:09:10.975 "raid_level": "raid1", 00:09:10.975 "superblock": true, 00:09:10.975 "num_base_bdevs": 2, 00:09:10.975 "num_base_bdevs_discovered": 2, 00:09:10.975 "num_base_bdevs_operational": 2, 00:09:10.975 "base_bdevs_list": [ 00:09:10.975 { 00:09:10.975 "name": "BaseBdev1", 00:09:10.975 "uuid": "cc90be56-ed87-572c-86ab-b3763dc22085", 00:09:10.975 "is_configured": true, 00:09:10.975 "data_offset": 2048, 00:09:10.975 "data_size": 63488 00:09:10.975 }, 00:09:10.975 { 00:09:10.975 "name": "BaseBdev2", 00:09:10.975 "uuid": "82b4866a-35c4-54de-801f-75f2bfc055d2", 00:09:10.975 "is_configured": true, 00:09:10.975 "data_offset": 2048, 00:09:10.975 "data_size": 63488 00:09:10.975 } 00:09:10.975 ] 00:09:10.975 }' 00:09:10.975 15:15:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:10.975 15:15:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.545 15:15:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:11.545 15:15:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.545 15:15:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.545 [2024-11-19 15:15:01.649362] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:11.545 [2024-11-19 15:15:01.649405] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:11.545 [2024-11-19 15:15:01.651996] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:11.545 [2024-11-19 15:15:01.652145] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:11.545 [2024-11-19 15:15:01.652269] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:11.545 [2024-11-19 15:15:01.652280] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:09:11.545 { 00:09:11.545 "results": [ 00:09:11.545 { 00:09:11.545 "job": "raid_bdev1", 00:09:11.545 "core_mask": "0x1", 00:09:11.545 "workload": "randrw", 00:09:11.545 "percentage": 50, 00:09:11.545 "status": "finished", 00:09:11.545 "queue_depth": 1, 00:09:11.545 "io_size": 131072, 00:09:11.545 "runtime": 1.391204, 00:09:11.545 "iops": 15154.499268259724, 00:09:11.545 "mibps": 1894.3124085324655, 00:09:11.545 "io_failed": 0, 00:09:11.545 "io_timeout": 0, 00:09:11.545 "avg_latency_us": 63.44720411548699, 00:09:11.545 "min_latency_us": 22.022707423580787, 00:09:11.545 "max_latency_us": 1459.5353711790392 00:09:11.545 } 00:09:11.545 ], 00:09:11.545 "core_count": 1 00:09:11.545 } 00:09:11.545 15:15:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.545 15:15:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 74763 00:09:11.545 15:15:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 74763 ']' 00:09:11.545 15:15:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 74763 00:09:11.545 15:15:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:09:11.545 15:15:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:11.545 15:15:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 74763 00:09:11.545 15:15:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:11.545 15:15:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:11.545 15:15:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 74763' 00:09:11.545 killing process with pid 74763 00:09:11.545 15:15:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 74763 00:09:11.545 [2024-11-19 15:15:01.698728] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:11.545 15:15:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 74763 00:09:11.545 [2024-11-19 15:15:01.727025] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:11.806 15:15:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:09:11.806 15:15:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:09:11.806 15:15:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.x1oRtBTOgW 00:09:11.806 ************************************ 00:09:11.806 END TEST raid_read_error_test 00:09:11.806 ************************************ 00:09:11.806 15:15:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:09:11.806 15:15:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:09:11.806 15:15:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:11.806 15:15:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:09:11.806 15:15:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:09:11.806 00:09:11.806 real 0m3.392s 00:09:11.806 user 0m4.243s 00:09:11.806 sys 0m0.554s 00:09:11.806 15:15:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:11.806 15:15:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.806 15:15:02 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid1 2 write 00:09:11.806 15:15:02 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:11.806 15:15:02 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:11.806 15:15:02 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:11.806 ************************************ 00:09:11.806 START TEST raid_write_error_test 00:09:11.806 ************************************ 00:09:11.806 15:15:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid1 2 write 00:09:11.806 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:09:11.806 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:09:11.806 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:09:11.806 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:09:11.806 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:11.806 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:09:11.806 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:11.806 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:11.806 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:09:11.806 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:11.806 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:11.806 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:09:11.806 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:09:11.806 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:09:11.806 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:09:11.806 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:09:11.806 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:09:11.806 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:09:11.806 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:09:11.806 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:09:11.806 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:09:11.806 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.8HFB7sC1rj 00:09:11.806 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=74897 00:09:11.806 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:09:11.806 15:15:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 74897 00:09:11.806 15:15:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 74897 ']' 00:09:12.067 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:12.067 15:15:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:12.067 15:15:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:12.067 15:15:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:12.067 15:15:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:12.067 15:15:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.067 [2024-11-19 15:15:02.218665] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:09:12.067 [2024-11-19 15:15:02.218805] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid74897 ] 00:09:12.067 [2024-11-19 15:15:02.376817] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:12.327 [2024-11-19 15:15:02.419680] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:12.327 [2024-11-19 15:15:02.496413] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:12.327 [2024-11-19 15:15:02.496458] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.897 BaseBdev1_malloc 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.897 true 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.897 [2024-11-19 15:15:03.142981] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:09:12.897 [2024-11-19 15:15:03.143041] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:12.897 [2024-11-19 15:15:03.143068] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:09:12.897 [2024-11-19 15:15:03.143079] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:12.897 [2024-11-19 15:15:03.145455] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:12.897 [2024-11-19 15:15:03.145492] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:09:12.897 BaseBdev1 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.897 BaseBdev2_malloc 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.897 true 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.897 [2024-11-19 15:15:03.189825] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:09:12.897 [2024-11-19 15:15:03.189896] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:12.897 [2024-11-19 15:15:03.189921] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:09:12.897 [2024-11-19 15:15:03.189941] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:12.897 [2024-11-19 15:15:03.192555] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:12.897 [2024-11-19 15:15:03.192597] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:09:12.897 BaseBdev2 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.897 [2024-11-19 15:15:03.201849] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:12.897 [2024-11-19 15:15:03.204134] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:12.897 [2024-11-19 15:15:03.204353] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:09:12.897 [2024-11-19 15:15:03.204376] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:12.897 [2024-11-19 15:15:03.204695] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:09:12.897 [2024-11-19 15:15:03.204860] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:09:12.897 [2024-11-19 15:15:03.204875] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:09:12.897 [2024-11-19 15:15:03.205109] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.897 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:13.157 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:13.157 "name": "raid_bdev1", 00:09:13.157 "uuid": "8bf3c34a-81a5-44e1-a57a-74d87937f463", 00:09:13.157 "strip_size_kb": 0, 00:09:13.157 "state": "online", 00:09:13.157 "raid_level": "raid1", 00:09:13.157 "superblock": true, 00:09:13.157 "num_base_bdevs": 2, 00:09:13.157 "num_base_bdevs_discovered": 2, 00:09:13.157 "num_base_bdevs_operational": 2, 00:09:13.157 "base_bdevs_list": [ 00:09:13.157 { 00:09:13.157 "name": "BaseBdev1", 00:09:13.157 "uuid": "aca18e10-5ae2-5343-938a-01b1a33c4da3", 00:09:13.157 "is_configured": true, 00:09:13.157 "data_offset": 2048, 00:09:13.157 "data_size": 63488 00:09:13.157 }, 00:09:13.157 { 00:09:13.157 "name": "BaseBdev2", 00:09:13.157 "uuid": "0dabfb44-3557-57d3-a96c-0a66745ab334", 00:09:13.157 "is_configured": true, 00:09:13.157 "data_offset": 2048, 00:09:13.157 "data_size": 63488 00:09:13.157 } 00:09:13.157 ] 00:09:13.157 }' 00:09:13.157 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:13.157 15:15:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:13.417 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:09:13.417 15:15:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:09:13.417 [2024-11-19 15:15:03.689563] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:09:14.356 15:15:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:09:14.356 15:15:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.356 15:15:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.356 [2024-11-19 15:15:04.612883] bdev_raid.c:2276:_raid_bdev_fail_base_bdev: *NOTICE*: Failing base bdev in slot 0 ('BaseBdev1') of raid bdev 'raid_bdev1' 00:09:14.356 [2024-11-19 15:15:04.612954] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:14.356 [2024-11-19 15:15:04.613210] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000002a10 00:09:14.356 15:15:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.356 15:15:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:09:14.356 15:15:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:09:14.356 15:15:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ write = \w\r\i\t\e ]] 00:09:14.356 15:15:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=1 00:09:14.356 15:15:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:09:14.356 15:15:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:14.356 15:15:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:14.356 15:15:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:14.356 15:15:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:14.357 15:15:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:09:14.357 15:15:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:14.357 15:15:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:14.357 15:15:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:14.357 15:15:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:14.357 15:15:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:14.357 15:15:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:14.357 15:15:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.357 15:15:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.357 15:15:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.357 15:15:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:14.357 "name": "raid_bdev1", 00:09:14.357 "uuid": "8bf3c34a-81a5-44e1-a57a-74d87937f463", 00:09:14.357 "strip_size_kb": 0, 00:09:14.357 "state": "online", 00:09:14.357 "raid_level": "raid1", 00:09:14.357 "superblock": true, 00:09:14.357 "num_base_bdevs": 2, 00:09:14.357 "num_base_bdevs_discovered": 1, 00:09:14.357 "num_base_bdevs_operational": 1, 00:09:14.357 "base_bdevs_list": [ 00:09:14.357 { 00:09:14.357 "name": null, 00:09:14.357 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:14.357 "is_configured": false, 00:09:14.357 "data_offset": 0, 00:09:14.357 "data_size": 63488 00:09:14.357 }, 00:09:14.357 { 00:09:14.357 "name": "BaseBdev2", 00:09:14.357 "uuid": "0dabfb44-3557-57d3-a96c-0a66745ab334", 00:09:14.357 "is_configured": true, 00:09:14.357 "data_offset": 2048, 00:09:14.357 "data_size": 63488 00:09:14.357 } 00:09:14.357 ] 00:09:14.357 }' 00:09:14.357 15:15:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:14.357 15:15:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.926 15:15:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:14.926 15:15:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.926 15:15:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.926 [2024-11-19 15:15:05.049989] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:14.926 [2024-11-19 15:15:05.050101] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:14.927 [2024-11-19 15:15:05.052692] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:14.927 [2024-11-19 15:15:05.052787] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:14.927 [2024-11-19 15:15:05.052861] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:14.927 [2024-11-19 15:15:05.052937] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:09:14.927 { 00:09:14.927 "results": [ 00:09:14.927 { 00:09:14.927 "job": "raid_bdev1", 00:09:14.927 "core_mask": "0x1", 00:09:14.927 "workload": "randrw", 00:09:14.927 "percentage": 50, 00:09:14.927 "status": "finished", 00:09:14.927 "queue_depth": 1, 00:09:14.927 "io_size": 131072, 00:09:14.927 "runtime": 1.360985, 00:09:14.927 "iops": 18434.442701425807, 00:09:14.927 "mibps": 2304.305337678226, 00:09:14.927 "io_failed": 0, 00:09:14.927 "io_timeout": 0, 00:09:14.927 "avg_latency_us": 51.68858225416208, 00:09:14.927 "min_latency_us": 22.246288209606988, 00:09:14.927 "max_latency_us": 1380.8349344978167 00:09:14.927 } 00:09:14.927 ], 00:09:14.927 "core_count": 1 00:09:14.927 } 00:09:14.927 15:15:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.927 15:15:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 74897 00:09:14.927 15:15:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 74897 ']' 00:09:14.927 15:15:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 74897 00:09:14.927 15:15:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:09:14.927 15:15:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:14.927 15:15:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 74897 00:09:14.927 15:15:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:14.927 15:15:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:14.927 killing process with pid 74897 00:09:14.927 15:15:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 74897' 00:09:14.927 15:15:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 74897 00:09:14.927 [2024-11-19 15:15:05.097649] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:14.927 15:15:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 74897 00:09:14.927 [2024-11-19 15:15:05.123510] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:15.187 15:15:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.8HFB7sC1rj 00:09:15.187 15:15:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:09:15.187 15:15:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:09:15.187 15:15:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:09:15.187 15:15:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:09:15.187 15:15:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:15.187 15:15:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:09:15.187 15:15:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:09:15.187 00:09:15.187 real 0m3.329s 00:09:15.187 user 0m4.148s 00:09:15.187 sys 0m0.566s 00:09:15.187 15:15:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:15.187 15:15:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.187 ************************************ 00:09:15.187 END TEST raid_write_error_test 00:09:15.187 ************************************ 00:09:15.187 15:15:05 bdev_raid -- bdev/bdev_raid.sh@966 -- # for n in {2..4} 00:09:15.187 15:15:05 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:09:15.187 15:15:05 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid0 3 false 00:09:15.187 15:15:05 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:15.187 15:15:05 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:15.187 15:15:05 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:15.187 ************************************ 00:09:15.187 START TEST raid_state_function_test 00:09:15.187 ************************************ 00:09:15.187 15:15:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid0 3 false 00:09:15.447 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:09:15.448 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:09:15.448 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:09:15.448 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:09:15.448 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:09:15.448 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:15.448 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:09:15.448 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:15.448 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:15.448 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:09:15.448 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:15.448 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:15.448 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:09:15.448 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:15.448 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:15.448 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:09:15.448 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:09:15.448 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:09:15.448 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:09:15.448 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:09:15.448 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:09:15.448 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:09:15.448 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:09:15.448 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:09:15.448 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:09:15.448 Process raid pid: 75024 00:09:15.448 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:09:15.448 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=75024 00:09:15.448 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:09:15.448 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 75024' 00:09:15.448 15:15:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 75024 00:09:15.448 15:15:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 75024 ']' 00:09:15.448 15:15:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:15.448 15:15:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:15.448 15:15:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:15.448 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:15.448 15:15:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:15.448 15:15:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.448 [2024-11-19 15:15:05.640463] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:09:15.448 [2024-11-19 15:15:05.640736] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:15.707 [2024-11-19 15:15:05.804643] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:15.707 [2024-11-19 15:15:05.844727] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:15.707 [2024-11-19 15:15:05.921048] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:15.707 [2024-11-19 15:15:05.921194] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:16.277 15:15:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:16.277 15:15:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:09:16.277 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:16.277 15:15:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:16.277 15:15:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.277 [2024-11-19 15:15:06.485198] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:16.277 [2024-11-19 15:15:06.485344] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:16.277 [2024-11-19 15:15:06.485378] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:16.277 [2024-11-19 15:15:06.485403] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:16.277 [2024-11-19 15:15:06.485421] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:16.277 [2024-11-19 15:15:06.485447] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:16.277 15:15:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:16.277 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:09:16.277 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:16.277 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:16.277 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:16.277 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:16.277 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:16.277 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:16.277 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:16.277 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:16.277 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:16.277 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:16.277 15:15:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:16.277 15:15:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.277 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:16.277 15:15:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:16.277 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:16.277 "name": "Existed_Raid", 00:09:16.277 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:16.277 "strip_size_kb": 64, 00:09:16.277 "state": "configuring", 00:09:16.277 "raid_level": "raid0", 00:09:16.277 "superblock": false, 00:09:16.277 "num_base_bdevs": 3, 00:09:16.277 "num_base_bdevs_discovered": 0, 00:09:16.277 "num_base_bdevs_operational": 3, 00:09:16.277 "base_bdevs_list": [ 00:09:16.277 { 00:09:16.277 "name": "BaseBdev1", 00:09:16.277 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:16.277 "is_configured": false, 00:09:16.277 "data_offset": 0, 00:09:16.277 "data_size": 0 00:09:16.277 }, 00:09:16.277 { 00:09:16.277 "name": "BaseBdev2", 00:09:16.277 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:16.277 "is_configured": false, 00:09:16.277 "data_offset": 0, 00:09:16.277 "data_size": 0 00:09:16.277 }, 00:09:16.277 { 00:09:16.277 "name": "BaseBdev3", 00:09:16.277 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:16.277 "is_configured": false, 00:09:16.277 "data_offset": 0, 00:09:16.277 "data_size": 0 00:09:16.277 } 00:09:16.277 ] 00:09:16.277 }' 00:09:16.277 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:16.277 15:15:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.845 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:16.845 15:15:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:16.845 15:15:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.845 [2024-11-19 15:15:06.912316] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:16.845 [2024-11-19 15:15:06.912371] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:09:16.845 15:15:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:16.845 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:16.845 15:15:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:16.845 15:15:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.845 [2024-11-19 15:15:06.924276] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:16.845 [2024-11-19 15:15:06.924321] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:16.845 [2024-11-19 15:15:06.924331] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:16.845 [2024-11-19 15:15:06.924341] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:16.845 [2024-11-19 15:15:06.924347] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:16.845 [2024-11-19 15:15:06.924357] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:16.845 15:15:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:16.845 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:16.845 15:15:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:16.845 15:15:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.845 [2024-11-19 15:15:06.951291] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:16.845 BaseBdev1 00:09:16.845 15:15:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:16.845 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:09:16.845 15:15:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:16.845 15:15:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:16.845 15:15:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:16.845 15:15:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:16.845 15:15:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:16.845 15:15:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:16.845 15:15:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:16.845 15:15:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.845 15:15:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:16.845 15:15:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:16.845 15:15:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:16.845 15:15:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.845 [ 00:09:16.845 { 00:09:16.845 "name": "BaseBdev1", 00:09:16.845 "aliases": [ 00:09:16.845 "9b5516b6-f521-432b-b5c9-a46730ec0c44" 00:09:16.845 ], 00:09:16.845 "product_name": "Malloc disk", 00:09:16.845 "block_size": 512, 00:09:16.845 "num_blocks": 65536, 00:09:16.845 "uuid": "9b5516b6-f521-432b-b5c9-a46730ec0c44", 00:09:16.845 "assigned_rate_limits": { 00:09:16.845 "rw_ios_per_sec": 0, 00:09:16.845 "rw_mbytes_per_sec": 0, 00:09:16.845 "r_mbytes_per_sec": 0, 00:09:16.845 "w_mbytes_per_sec": 0 00:09:16.845 }, 00:09:16.845 "claimed": true, 00:09:16.845 "claim_type": "exclusive_write", 00:09:16.845 "zoned": false, 00:09:16.845 "supported_io_types": { 00:09:16.845 "read": true, 00:09:16.845 "write": true, 00:09:16.845 "unmap": true, 00:09:16.845 "flush": true, 00:09:16.845 "reset": true, 00:09:16.845 "nvme_admin": false, 00:09:16.845 "nvme_io": false, 00:09:16.845 "nvme_io_md": false, 00:09:16.845 "write_zeroes": true, 00:09:16.845 "zcopy": true, 00:09:16.845 "get_zone_info": false, 00:09:16.845 "zone_management": false, 00:09:16.845 "zone_append": false, 00:09:16.845 "compare": false, 00:09:16.845 "compare_and_write": false, 00:09:16.845 "abort": true, 00:09:16.845 "seek_hole": false, 00:09:16.845 "seek_data": false, 00:09:16.845 "copy": true, 00:09:16.845 "nvme_iov_md": false 00:09:16.845 }, 00:09:16.845 "memory_domains": [ 00:09:16.845 { 00:09:16.845 "dma_device_id": "system", 00:09:16.845 "dma_device_type": 1 00:09:16.845 }, 00:09:16.845 { 00:09:16.845 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:16.845 "dma_device_type": 2 00:09:16.845 } 00:09:16.845 ], 00:09:16.845 "driver_specific": {} 00:09:16.845 } 00:09:16.845 ] 00:09:16.845 15:15:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:16.845 15:15:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:16.845 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:09:16.845 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:16.845 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:16.845 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:16.845 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:16.845 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:16.845 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:16.845 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:16.845 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:16.845 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:16.845 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:16.845 15:15:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:16.845 15:15:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:16.845 15:15:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.845 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:16.845 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:16.845 "name": "Existed_Raid", 00:09:16.845 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:16.845 "strip_size_kb": 64, 00:09:16.845 "state": "configuring", 00:09:16.845 "raid_level": "raid0", 00:09:16.845 "superblock": false, 00:09:16.845 "num_base_bdevs": 3, 00:09:16.845 "num_base_bdevs_discovered": 1, 00:09:16.845 "num_base_bdevs_operational": 3, 00:09:16.845 "base_bdevs_list": [ 00:09:16.845 { 00:09:16.845 "name": "BaseBdev1", 00:09:16.845 "uuid": "9b5516b6-f521-432b-b5c9-a46730ec0c44", 00:09:16.845 "is_configured": true, 00:09:16.845 "data_offset": 0, 00:09:16.845 "data_size": 65536 00:09:16.845 }, 00:09:16.845 { 00:09:16.845 "name": "BaseBdev2", 00:09:16.845 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:16.845 "is_configured": false, 00:09:16.845 "data_offset": 0, 00:09:16.845 "data_size": 0 00:09:16.845 }, 00:09:16.845 { 00:09:16.845 "name": "BaseBdev3", 00:09:16.845 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:16.845 "is_configured": false, 00:09:16.845 "data_offset": 0, 00:09:16.845 "data_size": 0 00:09:16.845 } 00:09:16.845 ] 00:09:16.845 }' 00:09:16.845 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:16.845 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.105 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:17.105 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:17.105 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.105 [2024-11-19 15:15:07.354700] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:17.105 [2024-11-19 15:15:07.354778] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:09:17.105 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:17.105 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:17.105 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:17.105 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.105 [2024-11-19 15:15:07.366701] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:17.105 [2024-11-19 15:15:07.369030] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:17.105 [2024-11-19 15:15:07.369145] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:17.105 [2024-11-19 15:15:07.369159] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:17.105 [2024-11-19 15:15:07.369170] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:17.105 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:17.105 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:09:17.105 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:17.105 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:09:17.105 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:17.105 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:17.105 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:17.105 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:17.105 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:17.105 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:17.105 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:17.105 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:17.105 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:17.105 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:17.105 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:17.105 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:17.105 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.105 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:17.105 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:17.105 "name": "Existed_Raid", 00:09:17.105 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:17.105 "strip_size_kb": 64, 00:09:17.105 "state": "configuring", 00:09:17.105 "raid_level": "raid0", 00:09:17.105 "superblock": false, 00:09:17.105 "num_base_bdevs": 3, 00:09:17.105 "num_base_bdevs_discovered": 1, 00:09:17.105 "num_base_bdevs_operational": 3, 00:09:17.105 "base_bdevs_list": [ 00:09:17.105 { 00:09:17.105 "name": "BaseBdev1", 00:09:17.105 "uuid": "9b5516b6-f521-432b-b5c9-a46730ec0c44", 00:09:17.105 "is_configured": true, 00:09:17.105 "data_offset": 0, 00:09:17.105 "data_size": 65536 00:09:17.105 }, 00:09:17.105 { 00:09:17.105 "name": "BaseBdev2", 00:09:17.105 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:17.105 "is_configured": false, 00:09:17.105 "data_offset": 0, 00:09:17.105 "data_size": 0 00:09:17.105 }, 00:09:17.105 { 00:09:17.105 "name": "BaseBdev3", 00:09:17.105 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:17.105 "is_configured": false, 00:09:17.105 "data_offset": 0, 00:09:17.105 "data_size": 0 00:09:17.105 } 00:09:17.105 ] 00:09:17.105 }' 00:09:17.105 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:17.105 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.674 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:17.674 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:17.674 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.674 [2024-11-19 15:15:07.786620] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:17.674 BaseBdev2 00:09:17.674 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:17.674 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:09:17.674 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:17.674 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:17.674 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:17.674 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:17.674 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:17.674 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:17.674 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:17.674 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.674 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:17.674 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:17.674 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:17.674 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.674 [ 00:09:17.674 { 00:09:17.674 "name": "BaseBdev2", 00:09:17.674 "aliases": [ 00:09:17.674 "97b8085f-3b31-400a-bd72-81400f37ca64" 00:09:17.674 ], 00:09:17.674 "product_name": "Malloc disk", 00:09:17.674 "block_size": 512, 00:09:17.674 "num_blocks": 65536, 00:09:17.674 "uuid": "97b8085f-3b31-400a-bd72-81400f37ca64", 00:09:17.674 "assigned_rate_limits": { 00:09:17.674 "rw_ios_per_sec": 0, 00:09:17.674 "rw_mbytes_per_sec": 0, 00:09:17.674 "r_mbytes_per_sec": 0, 00:09:17.674 "w_mbytes_per_sec": 0 00:09:17.674 }, 00:09:17.674 "claimed": true, 00:09:17.674 "claim_type": "exclusive_write", 00:09:17.674 "zoned": false, 00:09:17.674 "supported_io_types": { 00:09:17.674 "read": true, 00:09:17.674 "write": true, 00:09:17.674 "unmap": true, 00:09:17.674 "flush": true, 00:09:17.674 "reset": true, 00:09:17.674 "nvme_admin": false, 00:09:17.674 "nvme_io": false, 00:09:17.674 "nvme_io_md": false, 00:09:17.674 "write_zeroes": true, 00:09:17.674 "zcopy": true, 00:09:17.674 "get_zone_info": false, 00:09:17.674 "zone_management": false, 00:09:17.674 "zone_append": false, 00:09:17.674 "compare": false, 00:09:17.674 "compare_and_write": false, 00:09:17.674 "abort": true, 00:09:17.674 "seek_hole": false, 00:09:17.674 "seek_data": false, 00:09:17.674 "copy": true, 00:09:17.674 "nvme_iov_md": false 00:09:17.674 }, 00:09:17.674 "memory_domains": [ 00:09:17.674 { 00:09:17.674 "dma_device_id": "system", 00:09:17.674 "dma_device_type": 1 00:09:17.674 }, 00:09:17.674 { 00:09:17.674 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:17.674 "dma_device_type": 2 00:09:17.674 } 00:09:17.674 ], 00:09:17.674 "driver_specific": {} 00:09:17.674 } 00:09:17.674 ] 00:09:17.674 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:17.674 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:17.674 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:17.674 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:17.674 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:09:17.674 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:17.674 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:17.674 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:17.674 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:17.674 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:17.674 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:17.674 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:17.674 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:17.674 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:17.674 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:17.674 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:17.674 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:17.674 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.674 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:17.674 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:17.674 "name": "Existed_Raid", 00:09:17.674 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:17.674 "strip_size_kb": 64, 00:09:17.674 "state": "configuring", 00:09:17.674 "raid_level": "raid0", 00:09:17.674 "superblock": false, 00:09:17.674 "num_base_bdevs": 3, 00:09:17.674 "num_base_bdevs_discovered": 2, 00:09:17.674 "num_base_bdevs_operational": 3, 00:09:17.674 "base_bdevs_list": [ 00:09:17.674 { 00:09:17.674 "name": "BaseBdev1", 00:09:17.674 "uuid": "9b5516b6-f521-432b-b5c9-a46730ec0c44", 00:09:17.674 "is_configured": true, 00:09:17.674 "data_offset": 0, 00:09:17.674 "data_size": 65536 00:09:17.674 }, 00:09:17.674 { 00:09:17.674 "name": "BaseBdev2", 00:09:17.674 "uuid": "97b8085f-3b31-400a-bd72-81400f37ca64", 00:09:17.674 "is_configured": true, 00:09:17.674 "data_offset": 0, 00:09:17.674 "data_size": 65536 00:09:17.674 }, 00:09:17.674 { 00:09:17.674 "name": "BaseBdev3", 00:09:17.674 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:17.674 "is_configured": false, 00:09:17.674 "data_offset": 0, 00:09:17.674 "data_size": 0 00:09:17.674 } 00:09:17.674 ] 00:09:17.674 }' 00:09:17.674 15:15:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:17.674 15:15:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.934 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:17.934 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:17.934 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.934 [2024-11-19 15:15:08.246154] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:17.934 [2024-11-19 15:15:08.246205] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:09:17.934 [2024-11-19 15:15:08.246218] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:09:17.934 [2024-11-19 15:15:08.246525] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:09:17.934 [2024-11-19 15:15:08.246692] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:09:17.934 [2024-11-19 15:15:08.246707] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:09:17.934 [2024-11-19 15:15:08.246921] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:17.934 BaseBdev3 00:09:17.934 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:17.934 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:09:17.934 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:09:17.934 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:17.934 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:17.934 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:17.934 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:17.934 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:17.934 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:17.934 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.934 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:17.934 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:17.934 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:17.934 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.934 [ 00:09:17.934 { 00:09:17.934 "name": "BaseBdev3", 00:09:17.934 "aliases": [ 00:09:17.934 "3ba672e9-368b-4a6c-80e2-c12c737510b0" 00:09:17.934 ], 00:09:17.934 "product_name": "Malloc disk", 00:09:17.934 "block_size": 512, 00:09:17.934 "num_blocks": 65536, 00:09:17.934 "uuid": "3ba672e9-368b-4a6c-80e2-c12c737510b0", 00:09:17.934 "assigned_rate_limits": { 00:09:17.934 "rw_ios_per_sec": 0, 00:09:17.934 "rw_mbytes_per_sec": 0, 00:09:17.934 "r_mbytes_per_sec": 0, 00:09:17.934 "w_mbytes_per_sec": 0 00:09:17.934 }, 00:09:17.934 "claimed": true, 00:09:17.934 "claim_type": "exclusive_write", 00:09:17.934 "zoned": false, 00:09:17.934 "supported_io_types": { 00:09:17.934 "read": true, 00:09:17.934 "write": true, 00:09:17.934 "unmap": true, 00:09:17.934 "flush": true, 00:09:17.934 "reset": true, 00:09:18.193 "nvme_admin": false, 00:09:18.193 "nvme_io": false, 00:09:18.193 "nvme_io_md": false, 00:09:18.193 "write_zeroes": true, 00:09:18.193 "zcopy": true, 00:09:18.194 "get_zone_info": false, 00:09:18.194 "zone_management": false, 00:09:18.194 "zone_append": false, 00:09:18.194 "compare": false, 00:09:18.194 "compare_and_write": false, 00:09:18.194 "abort": true, 00:09:18.194 "seek_hole": false, 00:09:18.194 "seek_data": false, 00:09:18.194 "copy": true, 00:09:18.194 "nvme_iov_md": false 00:09:18.194 }, 00:09:18.194 "memory_domains": [ 00:09:18.194 { 00:09:18.194 "dma_device_id": "system", 00:09:18.194 "dma_device_type": 1 00:09:18.194 }, 00:09:18.194 { 00:09:18.194 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:18.194 "dma_device_type": 2 00:09:18.194 } 00:09:18.194 ], 00:09:18.194 "driver_specific": {} 00:09:18.194 } 00:09:18.194 ] 00:09:18.194 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.194 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:18.194 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:18.194 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:18.194 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:09:18.194 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:18.194 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:18.194 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:18.194 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:18.194 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:18.194 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:18.194 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:18.194 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:18.194 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:18.194 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:18.194 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.194 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.194 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:18.194 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.194 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:18.194 "name": "Existed_Raid", 00:09:18.194 "uuid": "55d974a9-2bb0-424e-9f1d-b3a1f0315199", 00:09:18.194 "strip_size_kb": 64, 00:09:18.194 "state": "online", 00:09:18.194 "raid_level": "raid0", 00:09:18.194 "superblock": false, 00:09:18.194 "num_base_bdevs": 3, 00:09:18.194 "num_base_bdevs_discovered": 3, 00:09:18.194 "num_base_bdevs_operational": 3, 00:09:18.194 "base_bdevs_list": [ 00:09:18.194 { 00:09:18.194 "name": "BaseBdev1", 00:09:18.194 "uuid": "9b5516b6-f521-432b-b5c9-a46730ec0c44", 00:09:18.194 "is_configured": true, 00:09:18.194 "data_offset": 0, 00:09:18.194 "data_size": 65536 00:09:18.194 }, 00:09:18.194 { 00:09:18.194 "name": "BaseBdev2", 00:09:18.194 "uuid": "97b8085f-3b31-400a-bd72-81400f37ca64", 00:09:18.194 "is_configured": true, 00:09:18.194 "data_offset": 0, 00:09:18.194 "data_size": 65536 00:09:18.194 }, 00:09:18.194 { 00:09:18.194 "name": "BaseBdev3", 00:09:18.194 "uuid": "3ba672e9-368b-4a6c-80e2-c12c737510b0", 00:09:18.194 "is_configured": true, 00:09:18.194 "data_offset": 0, 00:09:18.194 "data_size": 65536 00:09:18.194 } 00:09:18.194 ] 00:09:18.194 }' 00:09:18.194 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:18.194 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.467 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:09:18.467 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:18.467 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:18.467 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:18.467 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:18.467 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:18.467 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:18.467 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:18.467 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.467 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.467 [2024-11-19 15:15:08.697862] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:18.467 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.467 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:18.467 "name": "Existed_Raid", 00:09:18.467 "aliases": [ 00:09:18.467 "55d974a9-2bb0-424e-9f1d-b3a1f0315199" 00:09:18.467 ], 00:09:18.467 "product_name": "Raid Volume", 00:09:18.467 "block_size": 512, 00:09:18.467 "num_blocks": 196608, 00:09:18.467 "uuid": "55d974a9-2bb0-424e-9f1d-b3a1f0315199", 00:09:18.467 "assigned_rate_limits": { 00:09:18.467 "rw_ios_per_sec": 0, 00:09:18.467 "rw_mbytes_per_sec": 0, 00:09:18.467 "r_mbytes_per_sec": 0, 00:09:18.467 "w_mbytes_per_sec": 0 00:09:18.467 }, 00:09:18.467 "claimed": false, 00:09:18.467 "zoned": false, 00:09:18.467 "supported_io_types": { 00:09:18.467 "read": true, 00:09:18.467 "write": true, 00:09:18.467 "unmap": true, 00:09:18.467 "flush": true, 00:09:18.467 "reset": true, 00:09:18.467 "nvme_admin": false, 00:09:18.467 "nvme_io": false, 00:09:18.467 "nvme_io_md": false, 00:09:18.467 "write_zeroes": true, 00:09:18.467 "zcopy": false, 00:09:18.467 "get_zone_info": false, 00:09:18.467 "zone_management": false, 00:09:18.467 "zone_append": false, 00:09:18.467 "compare": false, 00:09:18.467 "compare_and_write": false, 00:09:18.467 "abort": false, 00:09:18.467 "seek_hole": false, 00:09:18.467 "seek_data": false, 00:09:18.467 "copy": false, 00:09:18.467 "nvme_iov_md": false 00:09:18.467 }, 00:09:18.467 "memory_domains": [ 00:09:18.467 { 00:09:18.467 "dma_device_id": "system", 00:09:18.467 "dma_device_type": 1 00:09:18.467 }, 00:09:18.467 { 00:09:18.467 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:18.467 "dma_device_type": 2 00:09:18.467 }, 00:09:18.467 { 00:09:18.467 "dma_device_id": "system", 00:09:18.467 "dma_device_type": 1 00:09:18.467 }, 00:09:18.467 { 00:09:18.467 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:18.467 "dma_device_type": 2 00:09:18.467 }, 00:09:18.467 { 00:09:18.467 "dma_device_id": "system", 00:09:18.467 "dma_device_type": 1 00:09:18.467 }, 00:09:18.467 { 00:09:18.467 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:18.467 "dma_device_type": 2 00:09:18.467 } 00:09:18.467 ], 00:09:18.467 "driver_specific": { 00:09:18.467 "raid": { 00:09:18.467 "uuid": "55d974a9-2bb0-424e-9f1d-b3a1f0315199", 00:09:18.467 "strip_size_kb": 64, 00:09:18.467 "state": "online", 00:09:18.467 "raid_level": "raid0", 00:09:18.467 "superblock": false, 00:09:18.467 "num_base_bdevs": 3, 00:09:18.467 "num_base_bdevs_discovered": 3, 00:09:18.467 "num_base_bdevs_operational": 3, 00:09:18.467 "base_bdevs_list": [ 00:09:18.467 { 00:09:18.467 "name": "BaseBdev1", 00:09:18.467 "uuid": "9b5516b6-f521-432b-b5c9-a46730ec0c44", 00:09:18.467 "is_configured": true, 00:09:18.467 "data_offset": 0, 00:09:18.467 "data_size": 65536 00:09:18.467 }, 00:09:18.467 { 00:09:18.467 "name": "BaseBdev2", 00:09:18.467 "uuid": "97b8085f-3b31-400a-bd72-81400f37ca64", 00:09:18.467 "is_configured": true, 00:09:18.467 "data_offset": 0, 00:09:18.467 "data_size": 65536 00:09:18.468 }, 00:09:18.468 { 00:09:18.468 "name": "BaseBdev3", 00:09:18.468 "uuid": "3ba672e9-368b-4a6c-80e2-c12c737510b0", 00:09:18.468 "is_configured": true, 00:09:18.468 "data_offset": 0, 00:09:18.468 "data_size": 65536 00:09:18.468 } 00:09:18.468 ] 00:09:18.468 } 00:09:18.468 } 00:09:18.468 }' 00:09:18.468 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:18.468 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:09:18.468 BaseBdev2 00:09:18.468 BaseBdev3' 00:09:18.468 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:18.468 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:18.468 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:18.468 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:18.468 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:09:18.468 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.468 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.729 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.729 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:18.730 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:18.730 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:18.730 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:18.730 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.730 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.730 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:18.730 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.730 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:18.730 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:18.730 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:18.730 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:18.730 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:18.730 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.730 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.730 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.730 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:18.730 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:18.730 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:18.730 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.730 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.730 [2024-11-19 15:15:08.909213] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:18.730 [2024-11-19 15:15:08.909271] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:18.730 [2024-11-19 15:15:08.909340] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:18.730 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.730 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:09:18.730 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:09:18.730 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:18.730 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:18.730 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:09:18.730 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 2 00:09:18.730 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:18.730 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:09:18.730 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:18.730 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:18.730 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:18.730 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:18.730 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:18.730 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:18.730 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:18.730 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:18.730 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:18.730 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.730 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.730 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.730 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:18.730 "name": "Existed_Raid", 00:09:18.730 "uuid": "55d974a9-2bb0-424e-9f1d-b3a1f0315199", 00:09:18.730 "strip_size_kb": 64, 00:09:18.730 "state": "offline", 00:09:18.730 "raid_level": "raid0", 00:09:18.730 "superblock": false, 00:09:18.730 "num_base_bdevs": 3, 00:09:18.730 "num_base_bdevs_discovered": 2, 00:09:18.730 "num_base_bdevs_operational": 2, 00:09:18.730 "base_bdevs_list": [ 00:09:18.730 { 00:09:18.730 "name": null, 00:09:18.730 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:18.730 "is_configured": false, 00:09:18.730 "data_offset": 0, 00:09:18.730 "data_size": 65536 00:09:18.730 }, 00:09:18.730 { 00:09:18.730 "name": "BaseBdev2", 00:09:18.730 "uuid": "97b8085f-3b31-400a-bd72-81400f37ca64", 00:09:18.730 "is_configured": true, 00:09:18.730 "data_offset": 0, 00:09:18.730 "data_size": 65536 00:09:18.730 }, 00:09:18.730 { 00:09:18.730 "name": "BaseBdev3", 00:09:18.730 "uuid": "3ba672e9-368b-4a6c-80e2-c12c737510b0", 00:09:18.730 "is_configured": true, 00:09:18.730 "data_offset": 0, 00:09:18.730 "data_size": 65536 00:09:18.730 } 00:09:18.730 ] 00:09:18.730 }' 00:09:18.730 15:15:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:18.730 15:15:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.021 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:09:19.021 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:19.021 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:19.021 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:19.021 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.021 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.021 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.021 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:19.021 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:19.021 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:09:19.021 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.021 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.021 [2024-11-19 15:15:09.333510] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:19.021 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.021 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:19.021 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:19.280 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:19.280 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:19.280 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.280 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.281 [2024-11-19 15:15:09.413472] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:19.281 [2024-11-19 15:15:09.413536] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.281 BaseBdev2 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.281 [ 00:09:19.281 { 00:09:19.281 "name": "BaseBdev2", 00:09:19.281 "aliases": [ 00:09:19.281 "a21d8be5-1a8f-4889-9199-1ec3adabf2a6" 00:09:19.281 ], 00:09:19.281 "product_name": "Malloc disk", 00:09:19.281 "block_size": 512, 00:09:19.281 "num_blocks": 65536, 00:09:19.281 "uuid": "a21d8be5-1a8f-4889-9199-1ec3adabf2a6", 00:09:19.281 "assigned_rate_limits": { 00:09:19.281 "rw_ios_per_sec": 0, 00:09:19.281 "rw_mbytes_per_sec": 0, 00:09:19.281 "r_mbytes_per_sec": 0, 00:09:19.281 "w_mbytes_per_sec": 0 00:09:19.281 }, 00:09:19.281 "claimed": false, 00:09:19.281 "zoned": false, 00:09:19.281 "supported_io_types": { 00:09:19.281 "read": true, 00:09:19.281 "write": true, 00:09:19.281 "unmap": true, 00:09:19.281 "flush": true, 00:09:19.281 "reset": true, 00:09:19.281 "nvme_admin": false, 00:09:19.281 "nvme_io": false, 00:09:19.281 "nvme_io_md": false, 00:09:19.281 "write_zeroes": true, 00:09:19.281 "zcopy": true, 00:09:19.281 "get_zone_info": false, 00:09:19.281 "zone_management": false, 00:09:19.281 "zone_append": false, 00:09:19.281 "compare": false, 00:09:19.281 "compare_and_write": false, 00:09:19.281 "abort": true, 00:09:19.281 "seek_hole": false, 00:09:19.281 "seek_data": false, 00:09:19.281 "copy": true, 00:09:19.281 "nvme_iov_md": false 00:09:19.281 }, 00:09:19.281 "memory_domains": [ 00:09:19.281 { 00:09:19.281 "dma_device_id": "system", 00:09:19.281 "dma_device_type": 1 00:09:19.281 }, 00:09:19.281 { 00:09:19.281 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:19.281 "dma_device_type": 2 00:09:19.281 } 00:09:19.281 ], 00:09:19.281 "driver_specific": {} 00:09:19.281 } 00:09:19.281 ] 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.281 BaseBdev3 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.281 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.281 [ 00:09:19.281 { 00:09:19.281 "name": "BaseBdev3", 00:09:19.281 "aliases": [ 00:09:19.281 "2a62c1d9-f677-43ca-80b8-0cfb0d62ba73" 00:09:19.281 ], 00:09:19.281 "product_name": "Malloc disk", 00:09:19.281 "block_size": 512, 00:09:19.281 "num_blocks": 65536, 00:09:19.281 "uuid": "2a62c1d9-f677-43ca-80b8-0cfb0d62ba73", 00:09:19.281 "assigned_rate_limits": { 00:09:19.281 "rw_ios_per_sec": 0, 00:09:19.281 "rw_mbytes_per_sec": 0, 00:09:19.281 "r_mbytes_per_sec": 0, 00:09:19.281 "w_mbytes_per_sec": 0 00:09:19.281 }, 00:09:19.281 "claimed": false, 00:09:19.281 "zoned": false, 00:09:19.281 "supported_io_types": { 00:09:19.281 "read": true, 00:09:19.281 "write": true, 00:09:19.281 "unmap": true, 00:09:19.281 "flush": true, 00:09:19.281 "reset": true, 00:09:19.281 "nvme_admin": false, 00:09:19.281 "nvme_io": false, 00:09:19.281 "nvme_io_md": false, 00:09:19.281 "write_zeroes": true, 00:09:19.281 "zcopy": true, 00:09:19.281 "get_zone_info": false, 00:09:19.281 "zone_management": false, 00:09:19.281 "zone_append": false, 00:09:19.281 "compare": false, 00:09:19.281 "compare_and_write": false, 00:09:19.281 "abort": true, 00:09:19.281 "seek_hole": false, 00:09:19.281 "seek_data": false, 00:09:19.281 "copy": true, 00:09:19.281 "nvme_iov_md": false 00:09:19.281 }, 00:09:19.281 "memory_domains": [ 00:09:19.281 { 00:09:19.281 "dma_device_id": "system", 00:09:19.281 "dma_device_type": 1 00:09:19.281 }, 00:09:19.281 { 00:09:19.281 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:19.281 "dma_device_type": 2 00:09:19.281 } 00:09:19.281 ], 00:09:19.281 "driver_specific": {} 00:09:19.281 } 00:09:19.281 ] 00:09:19.282 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.282 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:19.282 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:19.282 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:19.282 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:19.282 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.282 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.541 [2024-11-19 15:15:09.619189] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:19.541 [2024-11-19 15:15:09.619240] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:19.541 [2024-11-19 15:15:09.619284] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:19.541 [2024-11-19 15:15:09.621603] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:19.541 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.541 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:09:19.541 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:19.541 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:19.541 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:19.541 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:19.541 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:19.541 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:19.541 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:19.541 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:19.541 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:19.541 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:19.541 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.541 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.541 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:19.541 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.541 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:19.541 "name": "Existed_Raid", 00:09:19.541 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:19.541 "strip_size_kb": 64, 00:09:19.541 "state": "configuring", 00:09:19.541 "raid_level": "raid0", 00:09:19.541 "superblock": false, 00:09:19.541 "num_base_bdevs": 3, 00:09:19.541 "num_base_bdevs_discovered": 2, 00:09:19.541 "num_base_bdevs_operational": 3, 00:09:19.541 "base_bdevs_list": [ 00:09:19.541 { 00:09:19.541 "name": "BaseBdev1", 00:09:19.541 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:19.541 "is_configured": false, 00:09:19.541 "data_offset": 0, 00:09:19.541 "data_size": 0 00:09:19.541 }, 00:09:19.541 { 00:09:19.541 "name": "BaseBdev2", 00:09:19.541 "uuid": "a21d8be5-1a8f-4889-9199-1ec3adabf2a6", 00:09:19.541 "is_configured": true, 00:09:19.541 "data_offset": 0, 00:09:19.541 "data_size": 65536 00:09:19.541 }, 00:09:19.541 { 00:09:19.541 "name": "BaseBdev3", 00:09:19.541 "uuid": "2a62c1d9-f677-43ca-80b8-0cfb0d62ba73", 00:09:19.541 "is_configured": true, 00:09:19.541 "data_offset": 0, 00:09:19.541 "data_size": 65536 00:09:19.541 } 00:09:19.541 ] 00:09:19.541 }' 00:09:19.541 15:15:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:19.541 15:15:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.799 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:09:19.799 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.799 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.799 [2024-11-19 15:15:10.038531] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:19.799 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.799 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:09:19.799 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:19.799 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:19.799 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:19.799 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:19.799 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:19.799 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:19.799 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:19.799 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:19.799 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:19.799 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:19.799 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.799 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.799 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:19.799 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.799 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:19.799 "name": "Existed_Raid", 00:09:19.799 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:19.799 "strip_size_kb": 64, 00:09:19.800 "state": "configuring", 00:09:19.800 "raid_level": "raid0", 00:09:19.800 "superblock": false, 00:09:19.800 "num_base_bdevs": 3, 00:09:19.800 "num_base_bdevs_discovered": 1, 00:09:19.800 "num_base_bdevs_operational": 3, 00:09:19.800 "base_bdevs_list": [ 00:09:19.800 { 00:09:19.800 "name": "BaseBdev1", 00:09:19.800 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:19.800 "is_configured": false, 00:09:19.800 "data_offset": 0, 00:09:19.800 "data_size": 0 00:09:19.800 }, 00:09:19.800 { 00:09:19.800 "name": null, 00:09:19.800 "uuid": "a21d8be5-1a8f-4889-9199-1ec3adabf2a6", 00:09:19.800 "is_configured": false, 00:09:19.800 "data_offset": 0, 00:09:19.800 "data_size": 65536 00:09:19.800 }, 00:09:19.800 { 00:09:19.800 "name": "BaseBdev3", 00:09:19.800 "uuid": "2a62c1d9-f677-43ca-80b8-0cfb0d62ba73", 00:09:19.800 "is_configured": true, 00:09:19.800 "data_offset": 0, 00:09:19.800 "data_size": 65536 00:09:19.800 } 00:09:19.800 ] 00:09:19.800 }' 00:09:19.800 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:19.800 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.366 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:20.366 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:20.366 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.366 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.366 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.366 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:09:20.366 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:20.366 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.366 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.366 [2024-11-19 15:15:10.578588] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:20.366 BaseBdev1 00:09:20.366 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.366 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:09:20.366 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:20.366 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:20.366 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:20.366 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:20.366 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:20.366 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:20.366 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.366 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.366 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.366 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:20.366 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.366 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.366 [ 00:09:20.366 { 00:09:20.366 "name": "BaseBdev1", 00:09:20.366 "aliases": [ 00:09:20.366 "40baf02d-878e-4060-b7dc-59b73c941e7a" 00:09:20.366 ], 00:09:20.366 "product_name": "Malloc disk", 00:09:20.366 "block_size": 512, 00:09:20.366 "num_blocks": 65536, 00:09:20.366 "uuid": "40baf02d-878e-4060-b7dc-59b73c941e7a", 00:09:20.366 "assigned_rate_limits": { 00:09:20.366 "rw_ios_per_sec": 0, 00:09:20.366 "rw_mbytes_per_sec": 0, 00:09:20.366 "r_mbytes_per_sec": 0, 00:09:20.366 "w_mbytes_per_sec": 0 00:09:20.366 }, 00:09:20.366 "claimed": true, 00:09:20.366 "claim_type": "exclusive_write", 00:09:20.366 "zoned": false, 00:09:20.366 "supported_io_types": { 00:09:20.366 "read": true, 00:09:20.366 "write": true, 00:09:20.366 "unmap": true, 00:09:20.366 "flush": true, 00:09:20.366 "reset": true, 00:09:20.366 "nvme_admin": false, 00:09:20.366 "nvme_io": false, 00:09:20.366 "nvme_io_md": false, 00:09:20.366 "write_zeroes": true, 00:09:20.366 "zcopy": true, 00:09:20.366 "get_zone_info": false, 00:09:20.366 "zone_management": false, 00:09:20.366 "zone_append": false, 00:09:20.366 "compare": false, 00:09:20.366 "compare_and_write": false, 00:09:20.366 "abort": true, 00:09:20.366 "seek_hole": false, 00:09:20.366 "seek_data": false, 00:09:20.366 "copy": true, 00:09:20.366 "nvme_iov_md": false 00:09:20.366 }, 00:09:20.366 "memory_domains": [ 00:09:20.366 { 00:09:20.366 "dma_device_id": "system", 00:09:20.366 "dma_device_type": 1 00:09:20.366 }, 00:09:20.366 { 00:09:20.366 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:20.366 "dma_device_type": 2 00:09:20.366 } 00:09:20.366 ], 00:09:20.366 "driver_specific": {} 00:09:20.366 } 00:09:20.366 ] 00:09:20.366 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.366 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:20.366 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:09:20.366 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:20.366 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:20.366 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:20.366 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:20.366 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:20.366 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:20.366 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:20.366 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:20.366 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:20.366 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:20.366 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:20.366 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.366 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.367 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.367 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:20.367 "name": "Existed_Raid", 00:09:20.367 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:20.367 "strip_size_kb": 64, 00:09:20.367 "state": "configuring", 00:09:20.367 "raid_level": "raid0", 00:09:20.367 "superblock": false, 00:09:20.367 "num_base_bdevs": 3, 00:09:20.367 "num_base_bdevs_discovered": 2, 00:09:20.367 "num_base_bdevs_operational": 3, 00:09:20.367 "base_bdevs_list": [ 00:09:20.367 { 00:09:20.367 "name": "BaseBdev1", 00:09:20.367 "uuid": "40baf02d-878e-4060-b7dc-59b73c941e7a", 00:09:20.367 "is_configured": true, 00:09:20.367 "data_offset": 0, 00:09:20.367 "data_size": 65536 00:09:20.367 }, 00:09:20.367 { 00:09:20.367 "name": null, 00:09:20.367 "uuid": "a21d8be5-1a8f-4889-9199-1ec3adabf2a6", 00:09:20.367 "is_configured": false, 00:09:20.367 "data_offset": 0, 00:09:20.367 "data_size": 65536 00:09:20.367 }, 00:09:20.367 { 00:09:20.367 "name": "BaseBdev3", 00:09:20.367 "uuid": "2a62c1d9-f677-43ca-80b8-0cfb0d62ba73", 00:09:20.367 "is_configured": true, 00:09:20.367 "data_offset": 0, 00:09:20.367 "data_size": 65536 00:09:20.367 } 00:09:20.367 ] 00:09:20.367 }' 00:09:20.367 15:15:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:20.367 15:15:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.933 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:20.933 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.933 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.933 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:20.933 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.933 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:09:20.933 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:09:20.933 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.933 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.933 [2024-11-19 15:15:11.081849] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:20.933 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.933 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:09:20.933 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:20.933 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:20.933 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:20.933 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:20.934 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:20.934 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:20.934 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:20.934 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:20.934 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:20.934 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:20.934 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:20.934 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.934 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.934 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.934 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:20.934 "name": "Existed_Raid", 00:09:20.934 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:20.934 "strip_size_kb": 64, 00:09:20.934 "state": "configuring", 00:09:20.934 "raid_level": "raid0", 00:09:20.934 "superblock": false, 00:09:20.934 "num_base_bdevs": 3, 00:09:20.934 "num_base_bdevs_discovered": 1, 00:09:20.934 "num_base_bdevs_operational": 3, 00:09:20.934 "base_bdevs_list": [ 00:09:20.934 { 00:09:20.934 "name": "BaseBdev1", 00:09:20.934 "uuid": "40baf02d-878e-4060-b7dc-59b73c941e7a", 00:09:20.934 "is_configured": true, 00:09:20.934 "data_offset": 0, 00:09:20.934 "data_size": 65536 00:09:20.934 }, 00:09:20.934 { 00:09:20.934 "name": null, 00:09:20.934 "uuid": "a21d8be5-1a8f-4889-9199-1ec3adabf2a6", 00:09:20.934 "is_configured": false, 00:09:20.934 "data_offset": 0, 00:09:20.934 "data_size": 65536 00:09:20.934 }, 00:09:20.934 { 00:09:20.934 "name": null, 00:09:20.934 "uuid": "2a62c1d9-f677-43ca-80b8-0cfb0d62ba73", 00:09:20.934 "is_configured": false, 00:09:20.934 "data_offset": 0, 00:09:20.934 "data_size": 65536 00:09:20.934 } 00:09:20.934 ] 00:09:20.934 }' 00:09:20.934 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:20.934 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.195 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:21.195 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:21.195 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.195 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.195 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.195 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:09:21.195 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:09:21.195 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.195 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.195 [2024-11-19 15:15:11.513119] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:21.195 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.195 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:09:21.195 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:21.195 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:21.195 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:21.195 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:21.195 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:21.195 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:21.195 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:21.195 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:21.196 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:21.196 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:21.196 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.196 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.196 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:21.455 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.455 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:21.455 "name": "Existed_Raid", 00:09:21.455 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:21.455 "strip_size_kb": 64, 00:09:21.455 "state": "configuring", 00:09:21.455 "raid_level": "raid0", 00:09:21.455 "superblock": false, 00:09:21.455 "num_base_bdevs": 3, 00:09:21.455 "num_base_bdevs_discovered": 2, 00:09:21.455 "num_base_bdevs_operational": 3, 00:09:21.455 "base_bdevs_list": [ 00:09:21.455 { 00:09:21.455 "name": "BaseBdev1", 00:09:21.455 "uuid": "40baf02d-878e-4060-b7dc-59b73c941e7a", 00:09:21.455 "is_configured": true, 00:09:21.455 "data_offset": 0, 00:09:21.455 "data_size": 65536 00:09:21.455 }, 00:09:21.455 { 00:09:21.455 "name": null, 00:09:21.455 "uuid": "a21d8be5-1a8f-4889-9199-1ec3adabf2a6", 00:09:21.455 "is_configured": false, 00:09:21.455 "data_offset": 0, 00:09:21.455 "data_size": 65536 00:09:21.455 }, 00:09:21.455 { 00:09:21.455 "name": "BaseBdev3", 00:09:21.455 "uuid": "2a62c1d9-f677-43ca-80b8-0cfb0d62ba73", 00:09:21.455 "is_configured": true, 00:09:21.455 "data_offset": 0, 00:09:21.455 "data_size": 65536 00:09:21.455 } 00:09:21.455 ] 00:09:21.455 }' 00:09:21.455 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:21.455 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.712 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:21.712 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:21.713 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.713 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.713 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.713 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:09:21.713 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:21.713 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.713 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.713 [2024-11-19 15:15:11.972385] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:21.713 15:15:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.713 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:09:21.713 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:21.713 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:21.713 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:21.713 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:21.713 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:21.713 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:21.713 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:21.713 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:21.713 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:21.713 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:21.713 15:15:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:21.713 15:15:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.713 15:15:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.713 15:15:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.713 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:21.713 "name": "Existed_Raid", 00:09:21.713 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:21.713 "strip_size_kb": 64, 00:09:21.713 "state": "configuring", 00:09:21.713 "raid_level": "raid0", 00:09:21.713 "superblock": false, 00:09:21.713 "num_base_bdevs": 3, 00:09:21.713 "num_base_bdevs_discovered": 1, 00:09:21.713 "num_base_bdevs_operational": 3, 00:09:21.713 "base_bdevs_list": [ 00:09:21.713 { 00:09:21.713 "name": null, 00:09:21.713 "uuid": "40baf02d-878e-4060-b7dc-59b73c941e7a", 00:09:21.713 "is_configured": false, 00:09:21.713 "data_offset": 0, 00:09:21.713 "data_size": 65536 00:09:21.713 }, 00:09:21.713 { 00:09:21.713 "name": null, 00:09:21.713 "uuid": "a21d8be5-1a8f-4889-9199-1ec3adabf2a6", 00:09:21.713 "is_configured": false, 00:09:21.713 "data_offset": 0, 00:09:21.713 "data_size": 65536 00:09:21.713 }, 00:09:21.713 { 00:09:21.713 "name": "BaseBdev3", 00:09:21.713 "uuid": "2a62c1d9-f677-43ca-80b8-0cfb0d62ba73", 00:09:21.713 "is_configured": true, 00:09:21.713 "data_offset": 0, 00:09:21.713 "data_size": 65536 00:09:21.713 } 00:09:21.713 ] 00:09:21.713 }' 00:09:21.713 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:21.713 15:15:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.279 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:22.279 15:15:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.279 15:15:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.279 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:22.279 15:15:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.279 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:09:22.279 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:09:22.279 15:15:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.279 15:15:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.279 [2024-11-19 15:15:12.495458] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:22.279 15:15:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.279 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:09:22.279 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:22.279 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:22.279 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:22.279 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:22.280 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:22.280 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:22.280 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:22.280 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:22.280 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:22.280 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:22.280 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:22.280 15:15:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.280 15:15:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.280 15:15:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.280 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:22.280 "name": "Existed_Raid", 00:09:22.280 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:22.280 "strip_size_kb": 64, 00:09:22.280 "state": "configuring", 00:09:22.280 "raid_level": "raid0", 00:09:22.280 "superblock": false, 00:09:22.280 "num_base_bdevs": 3, 00:09:22.280 "num_base_bdevs_discovered": 2, 00:09:22.280 "num_base_bdevs_operational": 3, 00:09:22.280 "base_bdevs_list": [ 00:09:22.280 { 00:09:22.280 "name": null, 00:09:22.280 "uuid": "40baf02d-878e-4060-b7dc-59b73c941e7a", 00:09:22.280 "is_configured": false, 00:09:22.280 "data_offset": 0, 00:09:22.280 "data_size": 65536 00:09:22.280 }, 00:09:22.280 { 00:09:22.280 "name": "BaseBdev2", 00:09:22.280 "uuid": "a21d8be5-1a8f-4889-9199-1ec3adabf2a6", 00:09:22.280 "is_configured": true, 00:09:22.280 "data_offset": 0, 00:09:22.280 "data_size": 65536 00:09:22.280 }, 00:09:22.280 { 00:09:22.280 "name": "BaseBdev3", 00:09:22.280 "uuid": "2a62c1d9-f677-43ca-80b8-0cfb0d62ba73", 00:09:22.280 "is_configured": true, 00:09:22.280 "data_offset": 0, 00:09:22.280 "data_size": 65536 00:09:22.280 } 00:09:22.280 ] 00:09:22.280 }' 00:09:22.280 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:22.280 15:15:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.847 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:22.847 15:15:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.847 15:15:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.847 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:22.847 15:15:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.847 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:09:22.847 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:09:22.847 15:15:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:22.847 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.847 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.847 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.847 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 40baf02d-878e-4060-b7dc-59b73c941e7a 00:09:22.847 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.847 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.847 [2024-11-19 15:15:13.043359] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:09:22.847 [2024-11-19 15:15:13.043414] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:09:22.847 [2024-11-19 15:15:13.043426] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:09:22.847 [2024-11-19 15:15:13.043702] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:09:22.847 [2024-11-19 15:15:13.043880] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:09:22.847 [2024-11-19 15:15:13.043895] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:09:22.847 [2024-11-19 15:15:13.044119] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:22.847 NewBaseBdev 00:09:22.847 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.847 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:09:22.847 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:09:22.847 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:22.847 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:22.847 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:22.847 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:22.847 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:22.847 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.847 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.847 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.847 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:09:22.847 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.847 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.847 [ 00:09:22.847 { 00:09:22.847 "name": "NewBaseBdev", 00:09:22.847 "aliases": [ 00:09:22.847 "40baf02d-878e-4060-b7dc-59b73c941e7a" 00:09:22.847 ], 00:09:22.847 "product_name": "Malloc disk", 00:09:22.847 "block_size": 512, 00:09:22.847 "num_blocks": 65536, 00:09:22.847 "uuid": "40baf02d-878e-4060-b7dc-59b73c941e7a", 00:09:22.847 "assigned_rate_limits": { 00:09:22.847 "rw_ios_per_sec": 0, 00:09:22.847 "rw_mbytes_per_sec": 0, 00:09:22.847 "r_mbytes_per_sec": 0, 00:09:22.847 "w_mbytes_per_sec": 0 00:09:22.847 }, 00:09:22.847 "claimed": true, 00:09:22.847 "claim_type": "exclusive_write", 00:09:22.847 "zoned": false, 00:09:22.847 "supported_io_types": { 00:09:22.847 "read": true, 00:09:22.847 "write": true, 00:09:22.847 "unmap": true, 00:09:22.847 "flush": true, 00:09:22.847 "reset": true, 00:09:22.847 "nvme_admin": false, 00:09:22.847 "nvme_io": false, 00:09:22.847 "nvme_io_md": false, 00:09:22.847 "write_zeroes": true, 00:09:22.847 "zcopy": true, 00:09:22.847 "get_zone_info": false, 00:09:22.847 "zone_management": false, 00:09:22.847 "zone_append": false, 00:09:22.847 "compare": false, 00:09:22.847 "compare_and_write": false, 00:09:22.847 "abort": true, 00:09:22.847 "seek_hole": false, 00:09:22.847 "seek_data": false, 00:09:22.847 "copy": true, 00:09:22.847 "nvme_iov_md": false 00:09:22.847 }, 00:09:22.847 "memory_domains": [ 00:09:22.847 { 00:09:22.847 "dma_device_id": "system", 00:09:22.847 "dma_device_type": 1 00:09:22.847 }, 00:09:22.847 { 00:09:22.847 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:22.847 "dma_device_type": 2 00:09:22.847 } 00:09:22.847 ], 00:09:22.847 "driver_specific": {} 00:09:22.847 } 00:09:22.847 ] 00:09:22.847 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.847 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:22.847 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:09:22.847 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:22.847 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:22.847 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:22.847 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:22.847 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:22.847 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:22.847 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:22.847 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:22.847 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:22.847 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:22.847 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.847 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.847 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:22.847 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.847 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:22.847 "name": "Existed_Raid", 00:09:22.847 "uuid": "3c0aa122-b26b-4240-b204-4345d306d3e5", 00:09:22.847 "strip_size_kb": 64, 00:09:22.847 "state": "online", 00:09:22.847 "raid_level": "raid0", 00:09:22.847 "superblock": false, 00:09:22.847 "num_base_bdevs": 3, 00:09:22.847 "num_base_bdevs_discovered": 3, 00:09:22.847 "num_base_bdevs_operational": 3, 00:09:22.847 "base_bdevs_list": [ 00:09:22.847 { 00:09:22.847 "name": "NewBaseBdev", 00:09:22.847 "uuid": "40baf02d-878e-4060-b7dc-59b73c941e7a", 00:09:22.847 "is_configured": true, 00:09:22.847 "data_offset": 0, 00:09:22.847 "data_size": 65536 00:09:22.847 }, 00:09:22.847 { 00:09:22.847 "name": "BaseBdev2", 00:09:22.847 "uuid": "a21d8be5-1a8f-4889-9199-1ec3adabf2a6", 00:09:22.847 "is_configured": true, 00:09:22.847 "data_offset": 0, 00:09:22.847 "data_size": 65536 00:09:22.847 }, 00:09:22.847 { 00:09:22.847 "name": "BaseBdev3", 00:09:22.847 "uuid": "2a62c1d9-f677-43ca-80b8-0cfb0d62ba73", 00:09:22.847 "is_configured": true, 00:09:22.847 "data_offset": 0, 00:09:22.847 "data_size": 65536 00:09:22.847 } 00:09:22.847 ] 00:09:22.847 }' 00:09:22.847 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:22.847 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.415 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:09:23.415 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:23.415 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:23.415 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:23.415 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:23.415 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:23.415 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:23.415 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:23.415 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:23.415 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.415 [2024-11-19 15:15:13.554783] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:23.415 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:23.415 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:23.415 "name": "Existed_Raid", 00:09:23.415 "aliases": [ 00:09:23.415 "3c0aa122-b26b-4240-b204-4345d306d3e5" 00:09:23.415 ], 00:09:23.415 "product_name": "Raid Volume", 00:09:23.415 "block_size": 512, 00:09:23.415 "num_blocks": 196608, 00:09:23.415 "uuid": "3c0aa122-b26b-4240-b204-4345d306d3e5", 00:09:23.415 "assigned_rate_limits": { 00:09:23.415 "rw_ios_per_sec": 0, 00:09:23.415 "rw_mbytes_per_sec": 0, 00:09:23.415 "r_mbytes_per_sec": 0, 00:09:23.415 "w_mbytes_per_sec": 0 00:09:23.415 }, 00:09:23.415 "claimed": false, 00:09:23.415 "zoned": false, 00:09:23.415 "supported_io_types": { 00:09:23.415 "read": true, 00:09:23.415 "write": true, 00:09:23.415 "unmap": true, 00:09:23.415 "flush": true, 00:09:23.415 "reset": true, 00:09:23.415 "nvme_admin": false, 00:09:23.415 "nvme_io": false, 00:09:23.415 "nvme_io_md": false, 00:09:23.415 "write_zeroes": true, 00:09:23.415 "zcopy": false, 00:09:23.415 "get_zone_info": false, 00:09:23.415 "zone_management": false, 00:09:23.415 "zone_append": false, 00:09:23.415 "compare": false, 00:09:23.415 "compare_and_write": false, 00:09:23.415 "abort": false, 00:09:23.415 "seek_hole": false, 00:09:23.415 "seek_data": false, 00:09:23.415 "copy": false, 00:09:23.415 "nvme_iov_md": false 00:09:23.415 }, 00:09:23.415 "memory_domains": [ 00:09:23.415 { 00:09:23.415 "dma_device_id": "system", 00:09:23.415 "dma_device_type": 1 00:09:23.415 }, 00:09:23.415 { 00:09:23.415 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:23.415 "dma_device_type": 2 00:09:23.415 }, 00:09:23.415 { 00:09:23.415 "dma_device_id": "system", 00:09:23.415 "dma_device_type": 1 00:09:23.415 }, 00:09:23.415 { 00:09:23.415 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:23.415 "dma_device_type": 2 00:09:23.415 }, 00:09:23.415 { 00:09:23.415 "dma_device_id": "system", 00:09:23.415 "dma_device_type": 1 00:09:23.415 }, 00:09:23.415 { 00:09:23.415 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:23.415 "dma_device_type": 2 00:09:23.415 } 00:09:23.415 ], 00:09:23.415 "driver_specific": { 00:09:23.415 "raid": { 00:09:23.415 "uuid": "3c0aa122-b26b-4240-b204-4345d306d3e5", 00:09:23.415 "strip_size_kb": 64, 00:09:23.415 "state": "online", 00:09:23.415 "raid_level": "raid0", 00:09:23.415 "superblock": false, 00:09:23.415 "num_base_bdevs": 3, 00:09:23.415 "num_base_bdevs_discovered": 3, 00:09:23.415 "num_base_bdevs_operational": 3, 00:09:23.415 "base_bdevs_list": [ 00:09:23.415 { 00:09:23.415 "name": "NewBaseBdev", 00:09:23.415 "uuid": "40baf02d-878e-4060-b7dc-59b73c941e7a", 00:09:23.415 "is_configured": true, 00:09:23.415 "data_offset": 0, 00:09:23.415 "data_size": 65536 00:09:23.415 }, 00:09:23.415 { 00:09:23.415 "name": "BaseBdev2", 00:09:23.415 "uuid": "a21d8be5-1a8f-4889-9199-1ec3adabf2a6", 00:09:23.415 "is_configured": true, 00:09:23.415 "data_offset": 0, 00:09:23.415 "data_size": 65536 00:09:23.415 }, 00:09:23.415 { 00:09:23.415 "name": "BaseBdev3", 00:09:23.415 "uuid": "2a62c1d9-f677-43ca-80b8-0cfb0d62ba73", 00:09:23.415 "is_configured": true, 00:09:23.415 "data_offset": 0, 00:09:23.415 "data_size": 65536 00:09:23.415 } 00:09:23.415 ] 00:09:23.415 } 00:09:23.415 } 00:09:23.415 }' 00:09:23.415 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:23.415 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:09:23.415 BaseBdev2 00:09:23.415 BaseBdev3' 00:09:23.415 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:23.415 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:23.415 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:23.415 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:23.415 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:09:23.415 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:23.415 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.415 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:23.415 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:23.415 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:23.415 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:23.415 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:23.415 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:23.415 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:23.415 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.415 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:23.674 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:23.674 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:23.674 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:23.674 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:23.674 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:23.674 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:23.674 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.674 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:23.674 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:23.674 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:23.674 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:23.674 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:23.674 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.674 [2024-11-19 15:15:13.822168] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:23.674 [2024-11-19 15:15:13.822215] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:23.674 [2024-11-19 15:15:13.822328] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:23.674 [2024-11-19 15:15:13.822401] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:23.674 [2024-11-19 15:15:13.822418] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:09:23.674 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:23.674 15:15:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 75024 00:09:23.674 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 75024 ']' 00:09:23.674 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 75024 00:09:23.674 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:09:23.674 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:23.674 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 75024 00:09:23.674 killing process with pid 75024 00:09:23.674 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:23.674 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:23.674 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 75024' 00:09:23.674 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 75024 00:09:23.674 [2024-11-19 15:15:13.859648] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:23.674 15:15:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 75024 00:09:23.674 [2024-11-19 15:15:13.920543] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:23.934 ************************************ 00:09:23.934 END TEST raid_state_function_test 00:09:23.934 ************************************ 00:09:23.934 15:15:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:09:23.934 00:09:23.934 real 0m8.720s 00:09:23.934 user 0m14.579s 00:09:23.934 sys 0m1.832s 00:09:23.934 15:15:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:23.934 15:15:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.195 15:15:14 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid0 3 true 00:09:24.195 15:15:14 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:24.195 15:15:14 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:24.195 15:15:14 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:24.195 ************************************ 00:09:24.195 START TEST raid_state_function_test_sb 00:09:24.195 ************************************ 00:09:24.195 15:15:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid0 3 true 00:09:24.195 15:15:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:09:24.195 15:15:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:09:24.195 15:15:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:09:24.195 15:15:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:09:24.195 15:15:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:09:24.195 15:15:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:24.195 15:15:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:09:24.195 15:15:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:24.195 15:15:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:24.195 15:15:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:09:24.195 15:15:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:24.195 15:15:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:24.195 15:15:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:09:24.195 15:15:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:24.195 15:15:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:24.195 15:15:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:09:24.195 15:15:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:09:24.195 15:15:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:09:24.195 15:15:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:09:24.195 15:15:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:09:24.195 15:15:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:09:24.195 15:15:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:09:24.195 15:15:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:09:24.195 15:15:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:09:24.195 15:15:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:09:24.195 15:15:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:09:24.195 15:15:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=75629 00:09:24.195 15:15:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:09:24.195 15:15:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 75629' 00:09:24.195 Process raid pid: 75629 00:09:24.195 15:15:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 75629 00:09:24.195 15:15:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 75629 ']' 00:09:24.195 15:15:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:24.195 15:15:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:24.195 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:24.195 15:15:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:24.195 15:15:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:24.195 15:15:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:24.195 [2024-11-19 15:15:14.414162] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:09:24.195 [2024-11-19 15:15:14.414281] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:24.455 [2024-11-19 15:15:14.571691] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:24.455 [2024-11-19 15:15:14.609679] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:24.455 [2024-11-19 15:15:14.685917] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:24.455 [2024-11-19 15:15:14.686818] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:25.025 15:15:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:25.025 15:15:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:09:25.025 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:25.025 15:15:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.025 15:15:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:25.025 [2024-11-19 15:15:15.238126] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:25.025 [2024-11-19 15:15:15.238185] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:25.025 [2024-11-19 15:15:15.238195] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:25.025 [2024-11-19 15:15:15.238204] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:25.025 [2024-11-19 15:15:15.238210] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:25.025 [2024-11-19 15:15:15.238338] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:25.025 15:15:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.025 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:09:25.025 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:25.025 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:25.025 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:25.025 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:25.025 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:25.025 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:25.025 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:25.025 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:25.025 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:25.025 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:25.025 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:25.025 15:15:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.025 15:15:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:25.025 15:15:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.025 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:25.025 "name": "Existed_Raid", 00:09:25.025 "uuid": "672bed1c-a8b7-4388-9e04-1794a264422f", 00:09:25.025 "strip_size_kb": 64, 00:09:25.025 "state": "configuring", 00:09:25.025 "raid_level": "raid0", 00:09:25.025 "superblock": true, 00:09:25.025 "num_base_bdevs": 3, 00:09:25.025 "num_base_bdevs_discovered": 0, 00:09:25.025 "num_base_bdevs_operational": 3, 00:09:25.025 "base_bdevs_list": [ 00:09:25.025 { 00:09:25.025 "name": "BaseBdev1", 00:09:25.025 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:25.025 "is_configured": false, 00:09:25.025 "data_offset": 0, 00:09:25.025 "data_size": 0 00:09:25.025 }, 00:09:25.025 { 00:09:25.025 "name": "BaseBdev2", 00:09:25.025 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:25.025 "is_configured": false, 00:09:25.025 "data_offset": 0, 00:09:25.025 "data_size": 0 00:09:25.025 }, 00:09:25.025 { 00:09:25.025 "name": "BaseBdev3", 00:09:25.025 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:25.025 "is_configured": false, 00:09:25.025 "data_offset": 0, 00:09:25.025 "data_size": 0 00:09:25.025 } 00:09:25.025 ] 00:09:25.025 }' 00:09:25.025 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:25.025 15:15:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:25.595 [2024-11-19 15:15:15.709151] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:25.595 [2024-11-19 15:15:15.709201] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:25.595 [2024-11-19 15:15:15.721169] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:25.595 [2024-11-19 15:15:15.721207] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:25.595 [2024-11-19 15:15:15.721217] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:25.595 [2024-11-19 15:15:15.721228] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:25.595 [2024-11-19 15:15:15.721234] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:25.595 [2024-11-19 15:15:15.721243] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:25.595 [2024-11-19 15:15:15.748383] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:25.595 BaseBdev1 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:25.595 [ 00:09:25.595 { 00:09:25.595 "name": "BaseBdev1", 00:09:25.595 "aliases": [ 00:09:25.595 "dca36eca-8eb4-4688-8ab0-17a94f59ddde" 00:09:25.595 ], 00:09:25.595 "product_name": "Malloc disk", 00:09:25.595 "block_size": 512, 00:09:25.595 "num_blocks": 65536, 00:09:25.595 "uuid": "dca36eca-8eb4-4688-8ab0-17a94f59ddde", 00:09:25.595 "assigned_rate_limits": { 00:09:25.595 "rw_ios_per_sec": 0, 00:09:25.595 "rw_mbytes_per_sec": 0, 00:09:25.595 "r_mbytes_per_sec": 0, 00:09:25.595 "w_mbytes_per_sec": 0 00:09:25.595 }, 00:09:25.595 "claimed": true, 00:09:25.595 "claim_type": "exclusive_write", 00:09:25.595 "zoned": false, 00:09:25.595 "supported_io_types": { 00:09:25.595 "read": true, 00:09:25.595 "write": true, 00:09:25.595 "unmap": true, 00:09:25.595 "flush": true, 00:09:25.595 "reset": true, 00:09:25.595 "nvme_admin": false, 00:09:25.595 "nvme_io": false, 00:09:25.595 "nvme_io_md": false, 00:09:25.595 "write_zeroes": true, 00:09:25.595 "zcopy": true, 00:09:25.595 "get_zone_info": false, 00:09:25.595 "zone_management": false, 00:09:25.595 "zone_append": false, 00:09:25.595 "compare": false, 00:09:25.595 "compare_and_write": false, 00:09:25.595 "abort": true, 00:09:25.595 "seek_hole": false, 00:09:25.595 "seek_data": false, 00:09:25.595 "copy": true, 00:09:25.595 "nvme_iov_md": false 00:09:25.595 }, 00:09:25.595 "memory_domains": [ 00:09:25.595 { 00:09:25.595 "dma_device_id": "system", 00:09:25.595 "dma_device_type": 1 00:09:25.595 }, 00:09:25.595 { 00:09:25.595 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:25.595 "dma_device_type": 2 00:09:25.595 } 00:09:25.595 ], 00:09:25.595 "driver_specific": {} 00:09:25.595 } 00:09:25.595 ] 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:25.595 "name": "Existed_Raid", 00:09:25.595 "uuid": "63b53f6e-9f12-4c2b-9437-5af6de5a994b", 00:09:25.595 "strip_size_kb": 64, 00:09:25.595 "state": "configuring", 00:09:25.595 "raid_level": "raid0", 00:09:25.595 "superblock": true, 00:09:25.595 "num_base_bdevs": 3, 00:09:25.595 "num_base_bdevs_discovered": 1, 00:09:25.595 "num_base_bdevs_operational": 3, 00:09:25.595 "base_bdevs_list": [ 00:09:25.595 { 00:09:25.595 "name": "BaseBdev1", 00:09:25.595 "uuid": "dca36eca-8eb4-4688-8ab0-17a94f59ddde", 00:09:25.595 "is_configured": true, 00:09:25.595 "data_offset": 2048, 00:09:25.595 "data_size": 63488 00:09:25.595 }, 00:09:25.595 { 00:09:25.595 "name": "BaseBdev2", 00:09:25.595 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:25.595 "is_configured": false, 00:09:25.595 "data_offset": 0, 00:09:25.595 "data_size": 0 00:09:25.595 }, 00:09:25.595 { 00:09:25.595 "name": "BaseBdev3", 00:09:25.595 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:25.595 "is_configured": false, 00:09:25.595 "data_offset": 0, 00:09:25.595 "data_size": 0 00:09:25.595 } 00:09:25.595 ] 00:09:25.595 }' 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:25.595 15:15:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:26.195 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:26.195 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:26.195 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:26.195 [2024-11-19 15:15:16.227592] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:26.195 [2024-11-19 15:15:16.227660] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:09:26.195 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:26.195 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:26.195 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:26.195 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:26.195 [2024-11-19 15:15:16.239618] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:26.195 [2024-11-19 15:15:16.241816] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:26.195 [2024-11-19 15:15:16.241854] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:26.195 [2024-11-19 15:15:16.241864] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:26.195 [2024-11-19 15:15:16.241874] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:26.195 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:26.195 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:09:26.195 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:26.195 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:09:26.195 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:26.195 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:26.195 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:26.195 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:26.195 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:26.195 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:26.195 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:26.195 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:26.195 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:26.195 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:26.195 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:26.195 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:26.195 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:26.195 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:26.195 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:26.195 "name": "Existed_Raid", 00:09:26.195 "uuid": "ef7a3965-ad19-4fec-b7d7-6ee959673038", 00:09:26.195 "strip_size_kb": 64, 00:09:26.195 "state": "configuring", 00:09:26.195 "raid_level": "raid0", 00:09:26.195 "superblock": true, 00:09:26.195 "num_base_bdevs": 3, 00:09:26.195 "num_base_bdevs_discovered": 1, 00:09:26.195 "num_base_bdevs_operational": 3, 00:09:26.195 "base_bdevs_list": [ 00:09:26.195 { 00:09:26.195 "name": "BaseBdev1", 00:09:26.195 "uuid": "dca36eca-8eb4-4688-8ab0-17a94f59ddde", 00:09:26.195 "is_configured": true, 00:09:26.195 "data_offset": 2048, 00:09:26.195 "data_size": 63488 00:09:26.195 }, 00:09:26.195 { 00:09:26.195 "name": "BaseBdev2", 00:09:26.195 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:26.195 "is_configured": false, 00:09:26.195 "data_offset": 0, 00:09:26.195 "data_size": 0 00:09:26.195 }, 00:09:26.195 { 00:09:26.195 "name": "BaseBdev3", 00:09:26.195 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:26.195 "is_configured": false, 00:09:26.195 "data_offset": 0, 00:09:26.195 "data_size": 0 00:09:26.195 } 00:09:26.195 ] 00:09:26.195 }' 00:09:26.195 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:26.195 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:26.455 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:26.455 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:26.455 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:26.455 [2024-11-19 15:15:16.663436] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:26.455 BaseBdev2 00:09:26.455 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:26.455 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:09:26.455 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:26.455 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:26.455 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:26.455 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:26.455 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:26.455 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:26.455 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:26.455 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:26.455 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:26.455 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:26.455 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:26.455 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:26.455 [ 00:09:26.455 { 00:09:26.455 "name": "BaseBdev2", 00:09:26.455 "aliases": [ 00:09:26.455 "a4f73e17-5c88-4521-8607-45ce976e36a1" 00:09:26.455 ], 00:09:26.455 "product_name": "Malloc disk", 00:09:26.455 "block_size": 512, 00:09:26.455 "num_blocks": 65536, 00:09:26.455 "uuid": "a4f73e17-5c88-4521-8607-45ce976e36a1", 00:09:26.455 "assigned_rate_limits": { 00:09:26.455 "rw_ios_per_sec": 0, 00:09:26.455 "rw_mbytes_per_sec": 0, 00:09:26.455 "r_mbytes_per_sec": 0, 00:09:26.455 "w_mbytes_per_sec": 0 00:09:26.455 }, 00:09:26.455 "claimed": true, 00:09:26.455 "claim_type": "exclusive_write", 00:09:26.455 "zoned": false, 00:09:26.455 "supported_io_types": { 00:09:26.455 "read": true, 00:09:26.455 "write": true, 00:09:26.455 "unmap": true, 00:09:26.455 "flush": true, 00:09:26.455 "reset": true, 00:09:26.455 "nvme_admin": false, 00:09:26.455 "nvme_io": false, 00:09:26.455 "nvme_io_md": false, 00:09:26.455 "write_zeroes": true, 00:09:26.455 "zcopy": true, 00:09:26.455 "get_zone_info": false, 00:09:26.455 "zone_management": false, 00:09:26.455 "zone_append": false, 00:09:26.455 "compare": false, 00:09:26.455 "compare_and_write": false, 00:09:26.455 "abort": true, 00:09:26.455 "seek_hole": false, 00:09:26.455 "seek_data": false, 00:09:26.455 "copy": true, 00:09:26.455 "nvme_iov_md": false 00:09:26.455 }, 00:09:26.455 "memory_domains": [ 00:09:26.455 { 00:09:26.455 "dma_device_id": "system", 00:09:26.455 "dma_device_type": 1 00:09:26.455 }, 00:09:26.455 { 00:09:26.455 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:26.455 "dma_device_type": 2 00:09:26.455 } 00:09:26.455 ], 00:09:26.455 "driver_specific": {} 00:09:26.455 } 00:09:26.455 ] 00:09:26.455 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:26.455 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:26.455 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:26.455 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:26.455 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:09:26.455 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:26.455 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:26.455 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:26.455 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:26.455 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:26.455 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:26.455 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:26.455 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:26.455 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:26.455 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:26.456 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:26.456 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:26.456 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:26.456 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:26.456 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:26.456 "name": "Existed_Raid", 00:09:26.456 "uuid": "ef7a3965-ad19-4fec-b7d7-6ee959673038", 00:09:26.456 "strip_size_kb": 64, 00:09:26.456 "state": "configuring", 00:09:26.456 "raid_level": "raid0", 00:09:26.456 "superblock": true, 00:09:26.456 "num_base_bdevs": 3, 00:09:26.456 "num_base_bdevs_discovered": 2, 00:09:26.456 "num_base_bdevs_operational": 3, 00:09:26.456 "base_bdevs_list": [ 00:09:26.456 { 00:09:26.456 "name": "BaseBdev1", 00:09:26.456 "uuid": "dca36eca-8eb4-4688-8ab0-17a94f59ddde", 00:09:26.456 "is_configured": true, 00:09:26.456 "data_offset": 2048, 00:09:26.456 "data_size": 63488 00:09:26.456 }, 00:09:26.456 { 00:09:26.456 "name": "BaseBdev2", 00:09:26.456 "uuid": "a4f73e17-5c88-4521-8607-45ce976e36a1", 00:09:26.456 "is_configured": true, 00:09:26.456 "data_offset": 2048, 00:09:26.456 "data_size": 63488 00:09:26.456 }, 00:09:26.456 { 00:09:26.456 "name": "BaseBdev3", 00:09:26.456 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:26.456 "is_configured": false, 00:09:26.456 "data_offset": 0, 00:09:26.456 "data_size": 0 00:09:26.456 } 00:09:26.456 ] 00:09:26.456 }' 00:09:26.456 15:15:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:26.456 15:15:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:27.024 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:27.025 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.025 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:27.025 [2024-11-19 15:15:17.194409] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:27.025 [2024-11-19 15:15:17.194815] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:09:27.025 [2024-11-19 15:15:17.194886] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:09:27.025 BaseBdev3 00:09:27.025 [2024-11-19 15:15:17.195569] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:09:27.025 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.025 [2024-11-19 15:15:17.195898] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:09:27.025 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:09:27.025 [2024-11-19 15:15:17.195938] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:09:27.025 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:09:27.025 [2024-11-19 15:15:17.196285] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:27.025 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:27.025 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:27.025 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:27.025 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:27.025 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:27.025 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.025 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:27.025 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.025 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:27.025 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.025 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:27.025 [ 00:09:27.025 { 00:09:27.025 "name": "BaseBdev3", 00:09:27.025 "aliases": [ 00:09:27.025 "26955db9-6352-4377-a283-d608dac03eaa" 00:09:27.025 ], 00:09:27.025 "product_name": "Malloc disk", 00:09:27.025 "block_size": 512, 00:09:27.025 "num_blocks": 65536, 00:09:27.025 "uuid": "26955db9-6352-4377-a283-d608dac03eaa", 00:09:27.025 "assigned_rate_limits": { 00:09:27.025 "rw_ios_per_sec": 0, 00:09:27.025 "rw_mbytes_per_sec": 0, 00:09:27.025 "r_mbytes_per_sec": 0, 00:09:27.025 "w_mbytes_per_sec": 0 00:09:27.025 }, 00:09:27.025 "claimed": true, 00:09:27.025 "claim_type": "exclusive_write", 00:09:27.025 "zoned": false, 00:09:27.025 "supported_io_types": { 00:09:27.025 "read": true, 00:09:27.025 "write": true, 00:09:27.025 "unmap": true, 00:09:27.025 "flush": true, 00:09:27.025 "reset": true, 00:09:27.025 "nvme_admin": false, 00:09:27.025 "nvme_io": false, 00:09:27.025 "nvme_io_md": false, 00:09:27.025 "write_zeroes": true, 00:09:27.025 "zcopy": true, 00:09:27.025 "get_zone_info": false, 00:09:27.025 "zone_management": false, 00:09:27.025 "zone_append": false, 00:09:27.025 "compare": false, 00:09:27.025 "compare_and_write": false, 00:09:27.025 "abort": true, 00:09:27.025 "seek_hole": false, 00:09:27.025 "seek_data": false, 00:09:27.025 "copy": true, 00:09:27.025 "nvme_iov_md": false 00:09:27.025 }, 00:09:27.025 "memory_domains": [ 00:09:27.025 { 00:09:27.025 "dma_device_id": "system", 00:09:27.025 "dma_device_type": 1 00:09:27.025 }, 00:09:27.025 { 00:09:27.025 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:27.025 "dma_device_type": 2 00:09:27.025 } 00:09:27.025 ], 00:09:27.025 "driver_specific": {} 00:09:27.025 } 00:09:27.025 ] 00:09:27.025 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.025 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:27.025 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:27.025 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:27.025 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:09:27.025 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:27.025 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:27.025 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:27.025 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:27.025 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:27.025 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:27.025 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:27.025 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:27.025 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:27.025 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:27.025 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.025 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:27.025 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:27.025 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.025 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:27.025 "name": "Existed_Raid", 00:09:27.025 "uuid": "ef7a3965-ad19-4fec-b7d7-6ee959673038", 00:09:27.025 "strip_size_kb": 64, 00:09:27.025 "state": "online", 00:09:27.025 "raid_level": "raid0", 00:09:27.025 "superblock": true, 00:09:27.025 "num_base_bdevs": 3, 00:09:27.025 "num_base_bdevs_discovered": 3, 00:09:27.025 "num_base_bdevs_operational": 3, 00:09:27.025 "base_bdevs_list": [ 00:09:27.025 { 00:09:27.025 "name": "BaseBdev1", 00:09:27.025 "uuid": "dca36eca-8eb4-4688-8ab0-17a94f59ddde", 00:09:27.025 "is_configured": true, 00:09:27.025 "data_offset": 2048, 00:09:27.025 "data_size": 63488 00:09:27.025 }, 00:09:27.025 { 00:09:27.025 "name": "BaseBdev2", 00:09:27.025 "uuid": "a4f73e17-5c88-4521-8607-45ce976e36a1", 00:09:27.025 "is_configured": true, 00:09:27.025 "data_offset": 2048, 00:09:27.025 "data_size": 63488 00:09:27.025 }, 00:09:27.025 { 00:09:27.025 "name": "BaseBdev3", 00:09:27.025 "uuid": "26955db9-6352-4377-a283-d608dac03eaa", 00:09:27.025 "is_configured": true, 00:09:27.025 "data_offset": 2048, 00:09:27.025 "data_size": 63488 00:09:27.025 } 00:09:27.025 ] 00:09:27.025 }' 00:09:27.025 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:27.025 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:27.593 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:09:27.593 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:27.593 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:27.593 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:27.593 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:09:27.593 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:27.593 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:27.593 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:27.593 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.593 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:27.593 [2024-11-19 15:15:17.649848] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:27.593 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.593 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:27.593 "name": "Existed_Raid", 00:09:27.593 "aliases": [ 00:09:27.593 "ef7a3965-ad19-4fec-b7d7-6ee959673038" 00:09:27.593 ], 00:09:27.593 "product_name": "Raid Volume", 00:09:27.593 "block_size": 512, 00:09:27.593 "num_blocks": 190464, 00:09:27.593 "uuid": "ef7a3965-ad19-4fec-b7d7-6ee959673038", 00:09:27.593 "assigned_rate_limits": { 00:09:27.593 "rw_ios_per_sec": 0, 00:09:27.593 "rw_mbytes_per_sec": 0, 00:09:27.593 "r_mbytes_per_sec": 0, 00:09:27.593 "w_mbytes_per_sec": 0 00:09:27.593 }, 00:09:27.593 "claimed": false, 00:09:27.593 "zoned": false, 00:09:27.593 "supported_io_types": { 00:09:27.593 "read": true, 00:09:27.593 "write": true, 00:09:27.593 "unmap": true, 00:09:27.593 "flush": true, 00:09:27.593 "reset": true, 00:09:27.593 "nvme_admin": false, 00:09:27.593 "nvme_io": false, 00:09:27.593 "nvme_io_md": false, 00:09:27.593 "write_zeroes": true, 00:09:27.593 "zcopy": false, 00:09:27.593 "get_zone_info": false, 00:09:27.593 "zone_management": false, 00:09:27.593 "zone_append": false, 00:09:27.593 "compare": false, 00:09:27.593 "compare_and_write": false, 00:09:27.593 "abort": false, 00:09:27.593 "seek_hole": false, 00:09:27.593 "seek_data": false, 00:09:27.593 "copy": false, 00:09:27.593 "nvme_iov_md": false 00:09:27.593 }, 00:09:27.593 "memory_domains": [ 00:09:27.593 { 00:09:27.593 "dma_device_id": "system", 00:09:27.593 "dma_device_type": 1 00:09:27.593 }, 00:09:27.593 { 00:09:27.593 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:27.593 "dma_device_type": 2 00:09:27.593 }, 00:09:27.593 { 00:09:27.593 "dma_device_id": "system", 00:09:27.593 "dma_device_type": 1 00:09:27.593 }, 00:09:27.593 { 00:09:27.593 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:27.593 "dma_device_type": 2 00:09:27.593 }, 00:09:27.593 { 00:09:27.593 "dma_device_id": "system", 00:09:27.593 "dma_device_type": 1 00:09:27.593 }, 00:09:27.593 { 00:09:27.593 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:27.593 "dma_device_type": 2 00:09:27.593 } 00:09:27.593 ], 00:09:27.593 "driver_specific": { 00:09:27.593 "raid": { 00:09:27.593 "uuid": "ef7a3965-ad19-4fec-b7d7-6ee959673038", 00:09:27.593 "strip_size_kb": 64, 00:09:27.593 "state": "online", 00:09:27.593 "raid_level": "raid0", 00:09:27.593 "superblock": true, 00:09:27.593 "num_base_bdevs": 3, 00:09:27.593 "num_base_bdevs_discovered": 3, 00:09:27.593 "num_base_bdevs_operational": 3, 00:09:27.593 "base_bdevs_list": [ 00:09:27.593 { 00:09:27.593 "name": "BaseBdev1", 00:09:27.593 "uuid": "dca36eca-8eb4-4688-8ab0-17a94f59ddde", 00:09:27.593 "is_configured": true, 00:09:27.593 "data_offset": 2048, 00:09:27.593 "data_size": 63488 00:09:27.593 }, 00:09:27.593 { 00:09:27.593 "name": "BaseBdev2", 00:09:27.593 "uuid": "a4f73e17-5c88-4521-8607-45ce976e36a1", 00:09:27.593 "is_configured": true, 00:09:27.593 "data_offset": 2048, 00:09:27.593 "data_size": 63488 00:09:27.593 }, 00:09:27.593 { 00:09:27.593 "name": "BaseBdev3", 00:09:27.594 "uuid": "26955db9-6352-4377-a283-d608dac03eaa", 00:09:27.594 "is_configured": true, 00:09:27.594 "data_offset": 2048, 00:09:27.594 "data_size": 63488 00:09:27.594 } 00:09:27.594 ] 00:09:27.594 } 00:09:27.594 } 00:09:27.594 }' 00:09:27.594 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:27.594 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:09:27.594 BaseBdev2 00:09:27.594 BaseBdev3' 00:09:27.594 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:27.594 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:27.594 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:27.594 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:27.594 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:09:27.594 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.594 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:27.594 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.594 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:27.594 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:27.594 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:27.594 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:27.594 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:27.594 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.594 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:27.594 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.594 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:27.594 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:27.594 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:27.594 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:27.594 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:27.594 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.594 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:27.594 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.594 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:27.594 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:27.594 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:27.594 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.594 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:27.594 [2024-11-19 15:15:17.917142] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:27.594 [2024-11-19 15:15:17.917169] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:27.594 [2024-11-19 15:15:17.917236] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:27.854 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.854 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:09:27.854 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:09:27.854 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:27.854 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:09:27.854 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:09:27.854 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 2 00:09:27.854 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:27.854 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:09:27.854 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:27.854 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:27.854 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:27.854 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:27.854 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:27.854 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:27.854 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:27.854 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:27.854 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:27.854 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.854 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:27.854 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.854 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:27.854 "name": "Existed_Raid", 00:09:27.854 "uuid": "ef7a3965-ad19-4fec-b7d7-6ee959673038", 00:09:27.854 "strip_size_kb": 64, 00:09:27.854 "state": "offline", 00:09:27.854 "raid_level": "raid0", 00:09:27.854 "superblock": true, 00:09:27.854 "num_base_bdevs": 3, 00:09:27.854 "num_base_bdevs_discovered": 2, 00:09:27.854 "num_base_bdevs_operational": 2, 00:09:27.854 "base_bdevs_list": [ 00:09:27.854 { 00:09:27.854 "name": null, 00:09:27.854 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:27.854 "is_configured": false, 00:09:27.854 "data_offset": 0, 00:09:27.854 "data_size": 63488 00:09:27.854 }, 00:09:27.854 { 00:09:27.854 "name": "BaseBdev2", 00:09:27.854 "uuid": "a4f73e17-5c88-4521-8607-45ce976e36a1", 00:09:27.854 "is_configured": true, 00:09:27.854 "data_offset": 2048, 00:09:27.854 "data_size": 63488 00:09:27.854 }, 00:09:27.854 { 00:09:27.854 "name": "BaseBdev3", 00:09:27.854 "uuid": "26955db9-6352-4377-a283-d608dac03eaa", 00:09:27.854 "is_configured": true, 00:09:27.854 "data_offset": 2048, 00:09:27.854 "data_size": 63488 00:09:27.854 } 00:09:27.854 ] 00:09:27.854 }' 00:09:27.854 15:15:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:27.854 15:15:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:28.114 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:09:28.114 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:28.114 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:28.114 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.114 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:28.114 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:28.114 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.115 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:28.115 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:28.115 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:09:28.115 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.115 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:28.115 [2024-11-19 15:15:18.361436] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:28.115 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.115 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:28.115 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:28.115 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:28.115 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:28.115 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.115 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:28.115 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.115 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:28.115 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:28.115 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:09:28.115 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.115 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:28.115 [2024-11-19 15:15:18.429640] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:28.115 [2024-11-19 15:15:18.429794] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:09:28.376 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.376 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:28.376 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:28.376 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:28.376 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:09:28.376 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.376 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:28.376 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.376 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:09:28.376 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:09:28.376 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:09:28.376 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:09:28.376 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:28.376 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:28.376 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.376 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:28.376 BaseBdev2 00:09:28.376 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.376 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:09:28.376 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:28.376 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:28.376 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:28.376 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:28.376 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:28.376 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:28.376 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.376 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:28.376 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.376 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:28.376 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.376 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:28.376 [ 00:09:28.376 { 00:09:28.376 "name": "BaseBdev2", 00:09:28.376 "aliases": [ 00:09:28.376 "584debc2-8e92-4a6d-bfb2-005bcb6715fb" 00:09:28.376 ], 00:09:28.376 "product_name": "Malloc disk", 00:09:28.376 "block_size": 512, 00:09:28.376 "num_blocks": 65536, 00:09:28.376 "uuid": "584debc2-8e92-4a6d-bfb2-005bcb6715fb", 00:09:28.376 "assigned_rate_limits": { 00:09:28.376 "rw_ios_per_sec": 0, 00:09:28.376 "rw_mbytes_per_sec": 0, 00:09:28.376 "r_mbytes_per_sec": 0, 00:09:28.376 "w_mbytes_per_sec": 0 00:09:28.376 }, 00:09:28.376 "claimed": false, 00:09:28.376 "zoned": false, 00:09:28.376 "supported_io_types": { 00:09:28.376 "read": true, 00:09:28.376 "write": true, 00:09:28.376 "unmap": true, 00:09:28.376 "flush": true, 00:09:28.376 "reset": true, 00:09:28.376 "nvme_admin": false, 00:09:28.376 "nvme_io": false, 00:09:28.376 "nvme_io_md": false, 00:09:28.376 "write_zeroes": true, 00:09:28.376 "zcopy": true, 00:09:28.376 "get_zone_info": false, 00:09:28.376 "zone_management": false, 00:09:28.376 "zone_append": false, 00:09:28.376 "compare": false, 00:09:28.376 "compare_and_write": false, 00:09:28.376 "abort": true, 00:09:28.376 "seek_hole": false, 00:09:28.376 "seek_data": false, 00:09:28.376 "copy": true, 00:09:28.376 "nvme_iov_md": false 00:09:28.376 }, 00:09:28.376 "memory_domains": [ 00:09:28.376 { 00:09:28.376 "dma_device_id": "system", 00:09:28.376 "dma_device_type": 1 00:09:28.376 }, 00:09:28.376 { 00:09:28.376 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:28.376 "dma_device_type": 2 00:09:28.376 } 00:09:28.376 ], 00:09:28.377 "driver_specific": {} 00:09:28.377 } 00:09:28.377 ] 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:28.377 BaseBdev3 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:28.377 [ 00:09:28.377 { 00:09:28.377 "name": "BaseBdev3", 00:09:28.377 "aliases": [ 00:09:28.377 "1eacf62c-2995-4885-b370-0c7be5eb9f68" 00:09:28.377 ], 00:09:28.377 "product_name": "Malloc disk", 00:09:28.377 "block_size": 512, 00:09:28.377 "num_blocks": 65536, 00:09:28.377 "uuid": "1eacf62c-2995-4885-b370-0c7be5eb9f68", 00:09:28.377 "assigned_rate_limits": { 00:09:28.377 "rw_ios_per_sec": 0, 00:09:28.377 "rw_mbytes_per_sec": 0, 00:09:28.377 "r_mbytes_per_sec": 0, 00:09:28.377 "w_mbytes_per_sec": 0 00:09:28.377 }, 00:09:28.377 "claimed": false, 00:09:28.377 "zoned": false, 00:09:28.377 "supported_io_types": { 00:09:28.377 "read": true, 00:09:28.377 "write": true, 00:09:28.377 "unmap": true, 00:09:28.377 "flush": true, 00:09:28.377 "reset": true, 00:09:28.377 "nvme_admin": false, 00:09:28.377 "nvme_io": false, 00:09:28.377 "nvme_io_md": false, 00:09:28.377 "write_zeroes": true, 00:09:28.377 "zcopy": true, 00:09:28.377 "get_zone_info": false, 00:09:28.377 "zone_management": false, 00:09:28.377 "zone_append": false, 00:09:28.377 "compare": false, 00:09:28.377 "compare_and_write": false, 00:09:28.377 "abort": true, 00:09:28.377 "seek_hole": false, 00:09:28.377 "seek_data": false, 00:09:28.377 "copy": true, 00:09:28.377 "nvme_iov_md": false 00:09:28.377 }, 00:09:28.377 "memory_domains": [ 00:09:28.377 { 00:09:28.377 "dma_device_id": "system", 00:09:28.377 "dma_device_type": 1 00:09:28.377 }, 00:09:28.377 { 00:09:28.377 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:28.377 "dma_device_type": 2 00:09:28.377 } 00:09:28.377 ], 00:09:28.377 "driver_specific": {} 00:09:28.377 } 00:09:28.377 ] 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:28.377 [2024-11-19 15:15:18.620361] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:28.377 [2024-11-19 15:15:18.620503] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:28.377 [2024-11-19 15:15:18.620558] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:28.377 [2024-11-19 15:15:18.622767] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:28.377 "name": "Existed_Raid", 00:09:28.377 "uuid": "6c39dcfc-691f-4ca5-ad3f-a1438f4b24e7", 00:09:28.377 "strip_size_kb": 64, 00:09:28.377 "state": "configuring", 00:09:28.377 "raid_level": "raid0", 00:09:28.377 "superblock": true, 00:09:28.377 "num_base_bdevs": 3, 00:09:28.377 "num_base_bdevs_discovered": 2, 00:09:28.377 "num_base_bdevs_operational": 3, 00:09:28.377 "base_bdevs_list": [ 00:09:28.377 { 00:09:28.377 "name": "BaseBdev1", 00:09:28.377 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:28.377 "is_configured": false, 00:09:28.377 "data_offset": 0, 00:09:28.377 "data_size": 0 00:09:28.377 }, 00:09:28.377 { 00:09:28.377 "name": "BaseBdev2", 00:09:28.377 "uuid": "584debc2-8e92-4a6d-bfb2-005bcb6715fb", 00:09:28.377 "is_configured": true, 00:09:28.377 "data_offset": 2048, 00:09:28.377 "data_size": 63488 00:09:28.377 }, 00:09:28.377 { 00:09:28.377 "name": "BaseBdev3", 00:09:28.377 "uuid": "1eacf62c-2995-4885-b370-0c7be5eb9f68", 00:09:28.377 "is_configured": true, 00:09:28.377 "data_offset": 2048, 00:09:28.377 "data_size": 63488 00:09:28.377 } 00:09:28.377 ] 00:09:28.377 }' 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:28.377 15:15:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:28.946 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:09:28.946 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.946 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:28.946 [2024-11-19 15:15:19.083579] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:28.946 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.946 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:09:28.946 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:28.946 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:28.946 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:28.946 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:28.946 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:28.946 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:28.946 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:28.946 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:28.946 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:28.946 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:28.946 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:28.946 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.946 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:28.946 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.946 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:28.946 "name": "Existed_Raid", 00:09:28.946 "uuid": "6c39dcfc-691f-4ca5-ad3f-a1438f4b24e7", 00:09:28.946 "strip_size_kb": 64, 00:09:28.946 "state": "configuring", 00:09:28.946 "raid_level": "raid0", 00:09:28.946 "superblock": true, 00:09:28.946 "num_base_bdevs": 3, 00:09:28.946 "num_base_bdevs_discovered": 1, 00:09:28.946 "num_base_bdevs_operational": 3, 00:09:28.946 "base_bdevs_list": [ 00:09:28.946 { 00:09:28.946 "name": "BaseBdev1", 00:09:28.946 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:28.946 "is_configured": false, 00:09:28.946 "data_offset": 0, 00:09:28.946 "data_size": 0 00:09:28.946 }, 00:09:28.946 { 00:09:28.946 "name": null, 00:09:28.946 "uuid": "584debc2-8e92-4a6d-bfb2-005bcb6715fb", 00:09:28.946 "is_configured": false, 00:09:28.946 "data_offset": 0, 00:09:28.946 "data_size": 63488 00:09:28.946 }, 00:09:28.946 { 00:09:28.946 "name": "BaseBdev3", 00:09:28.946 "uuid": "1eacf62c-2995-4885-b370-0c7be5eb9f68", 00:09:28.946 "is_configured": true, 00:09:28.946 "data_offset": 2048, 00:09:28.946 "data_size": 63488 00:09:28.946 } 00:09:28.946 ] 00:09:28.946 }' 00:09:28.946 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:28.946 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:29.206 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:29.206 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:29.206 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.206 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:29.206 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.466 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:09:29.466 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:29.466 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.466 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:29.466 [2024-11-19 15:15:19.587708] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:29.466 BaseBdev1 00:09:29.466 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.466 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:09:29.466 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:29.466 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:29.466 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:29.466 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:29.466 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:29.466 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:29.466 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.466 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:29.466 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.466 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:29.466 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.466 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:29.466 [ 00:09:29.466 { 00:09:29.466 "name": "BaseBdev1", 00:09:29.466 "aliases": [ 00:09:29.466 "e83c6c66-9558-4d49-80fe-9da46fa19fa7" 00:09:29.466 ], 00:09:29.466 "product_name": "Malloc disk", 00:09:29.466 "block_size": 512, 00:09:29.466 "num_blocks": 65536, 00:09:29.466 "uuid": "e83c6c66-9558-4d49-80fe-9da46fa19fa7", 00:09:29.466 "assigned_rate_limits": { 00:09:29.466 "rw_ios_per_sec": 0, 00:09:29.466 "rw_mbytes_per_sec": 0, 00:09:29.466 "r_mbytes_per_sec": 0, 00:09:29.466 "w_mbytes_per_sec": 0 00:09:29.466 }, 00:09:29.466 "claimed": true, 00:09:29.466 "claim_type": "exclusive_write", 00:09:29.466 "zoned": false, 00:09:29.466 "supported_io_types": { 00:09:29.466 "read": true, 00:09:29.466 "write": true, 00:09:29.466 "unmap": true, 00:09:29.466 "flush": true, 00:09:29.466 "reset": true, 00:09:29.466 "nvme_admin": false, 00:09:29.466 "nvme_io": false, 00:09:29.466 "nvme_io_md": false, 00:09:29.466 "write_zeroes": true, 00:09:29.466 "zcopy": true, 00:09:29.466 "get_zone_info": false, 00:09:29.466 "zone_management": false, 00:09:29.466 "zone_append": false, 00:09:29.466 "compare": false, 00:09:29.466 "compare_and_write": false, 00:09:29.466 "abort": true, 00:09:29.466 "seek_hole": false, 00:09:29.466 "seek_data": false, 00:09:29.466 "copy": true, 00:09:29.466 "nvme_iov_md": false 00:09:29.466 }, 00:09:29.466 "memory_domains": [ 00:09:29.466 { 00:09:29.466 "dma_device_id": "system", 00:09:29.466 "dma_device_type": 1 00:09:29.466 }, 00:09:29.466 { 00:09:29.466 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:29.466 "dma_device_type": 2 00:09:29.466 } 00:09:29.466 ], 00:09:29.466 "driver_specific": {} 00:09:29.466 } 00:09:29.466 ] 00:09:29.466 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.466 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:29.466 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:09:29.466 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:29.466 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:29.466 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:29.466 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:29.466 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:29.466 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:29.466 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:29.466 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:29.466 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:29.467 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:29.467 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:29.467 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.467 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:29.467 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.467 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:29.467 "name": "Existed_Raid", 00:09:29.467 "uuid": "6c39dcfc-691f-4ca5-ad3f-a1438f4b24e7", 00:09:29.467 "strip_size_kb": 64, 00:09:29.467 "state": "configuring", 00:09:29.467 "raid_level": "raid0", 00:09:29.467 "superblock": true, 00:09:29.467 "num_base_bdevs": 3, 00:09:29.467 "num_base_bdevs_discovered": 2, 00:09:29.467 "num_base_bdevs_operational": 3, 00:09:29.467 "base_bdevs_list": [ 00:09:29.467 { 00:09:29.467 "name": "BaseBdev1", 00:09:29.467 "uuid": "e83c6c66-9558-4d49-80fe-9da46fa19fa7", 00:09:29.467 "is_configured": true, 00:09:29.467 "data_offset": 2048, 00:09:29.467 "data_size": 63488 00:09:29.467 }, 00:09:29.467 { 00:09:29.467 "name": null, 00:09:29.467 "uuid": "584debc2-8e92-4a6d-bfb2-005bcb6715fb", 00:09:29.467 "is_configured": false, 00:09:29.467 "data_offset": 0, 00:09:29.467 "data_size": 63488 00:09:29.467 }, 00:09:29.467 { 00:09:29.467 "name": "BaseBdev3", 00:09:29.467 "uuid": "1eacf62c-2995-4885-b370-0c7be5eb9f68", 00:09:29.467 "is_configured": true, 00:09:29.467 "data_offset": 2048, 00:09:29.467 "data_size": 63488 00:09:29.467 } 00:09:29.467 ] 00:09:29.467 }' 00:09:29.467 15:15:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:29.467 15:15:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:29.727 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:29.727 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:29.727 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.727 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:29.986 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.986 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:09:29.986 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:09:29.987 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.987 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:29.987 [2024-11-19 15:15:20.110870] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:29.987 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.987 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:09:29.987 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:29.987 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:29.987 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:29.987 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:29.987 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:29.987 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:29.987 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:29.987 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:29.987 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:29.987 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:29.987 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:29.987 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.987 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:29.987 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.987 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:29.987 "name": "Existed_Raid", 00:09:29.987 "uuid": "6c39dcfc-691f-4ca5-ad3f-a1438f4b24e7", 00:09:29.987 "strip_size_kb": 64, 00:09:29.987 "state": "configuring", 00:09:29.987 "raid_level": "raid0", 00:09:29.987 "superblock": true, 00:09:29.987 "num_base_bdevs": 3, 00:09:29.987 "num_base_bdevs_discovered": 1, 00:09:29.987 "num_base_bdevs_operational": 3, 00:09:29.987 "base_bdevs_list": [ 00:09:29.987 { 00:09:29.987 "name": "BaseBdev1", 00:09:29.987 "uuid": "e83c6c66-9558-4d49-80fe-9da46fa19fa7", 00:09:29.987 "is_configured": true, 00:09:29.987 "data_offset": 2048, 00:09:29.987 "data_size": 63488 00:09:29.987 }, 00:09:29.987 { 00:09:29.987 "name": null, 00:09:29.987 "uuid": "584debc2-8e92-4a6d-bfb2-005bcb6715fb", 00:09:29.987 "is_configured": false, 00:09:29.987 "data_offset": 0, 00:09:29.987 "data_size": 63488 00:09:29.987 }, 00:09:29.987 { 00:09:29.987 "name": null, 00:09:29.987 "uuid": "1eacf62c-2995-4885-b370-0c7be5eb9f68", 00:09:29.987 "is_configured": false, 00:09:29.987 "data_offset": 0, 00:09:29.987 "data_size": 63488 00:09:29.987 } 00:09:29.987 ] 00:09:29.987 }' 00:09:29.987 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:29.987 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.246 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:30.246 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.246 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.246 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:30.246 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.506 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:09:30.506 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:09:30.506 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.506 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.506 [2024-11-19 15:15:20.606098] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:30.506 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.506 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:09:30.506 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:30.506 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:30.506 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:30.506 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:30.506 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:30.506 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:30.506 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:30.506 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:30.506 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:30.506 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:30.506 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.506 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.506 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:30.506 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.507 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:30.507 "name": "Existed_Raid", 00:09:30.507 "uuid": "6c39dcfc-691f-4ca5-ad3f-a1438f4b24e7", 00:09:30.507 "strip_size_kb": 64, 00:09:30.507 "state": "configuring", 00:09:30.507 "raid_level": "raid0", 00:09:30.507 "superblock": true, 00:09:30.507 "num_base_bdevs": 3, 00:09:30.507 "num_base_bdevs_discovered": 2, 00:09:30.507 "num_base_bdevs_operational": 3, 00:09:30.507 "base_bdevs_list": [ 00:09:30.507 { 00:09:30.507 "name": "BaseBdev1", 00:09:30.507 "uuid": "e83c6c66-9558-4d49-80fe-9da46fa19fa7", 00:09:30.507 "is_configured": true, 00:09:30.507 "data_offset": 2048, 00:09:30.507 "data_size": 63488 00:09:30.507 }, 00:09:30.507 { 00:09:30.507 "name": null, 00:09:30.507 "uuid": "584debc2-8e92-4a6d-bfb2-005bcb6715fb", 00:09:30.507 "is_configured": false, 00:09:30.507 "data_offset": 0, 00:09:30.507 "data_size": 63488 00:09:30.507 }, 00:09:30.507 { 00:09:30.507 "name": "BaseBdev3", 00:09:30.507 "uuid": "1eacf62c-2995-4885-b370-0c7be5eb9f68", 00:09:30.507 "is_configured": true, 00:09:30.507 "data_offset": 2048, 00:09:30.507 "data_size": 63488 00:09:30.507 } 00:09:30.507 ] 00:09:30.507 }' 00:09:30.507 15:15:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:30.507 15:15:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.767 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:30.767 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:30.767 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.767 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.767 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.767 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:09:30.767 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:30.767 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.767 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.767 [2024-11-19 15:15:21.057348] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:30.767 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.767 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:09:30.767 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:30.767 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:30.767 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:30.767 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:30.767 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:30.767 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:30.767 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:30.767 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:30.767 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:30.767 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:30.768 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.768 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.768 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:30.768 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:31.027 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:31.027 "name": "Existed_Raid", 00:09:31.027 "uuid": "6c39dcfc-691f-4ca5-ad3f-a1438f4b24e7", 00:09:31.027 "strip_size_kb": 64, 00:09:31.027 "state": "configuring", 00:09:31.027 "raid_level": "raid0", 00:09:31.027 "superblock": true, 00:09:31.027 "num_base_bdevs": 3, 00:09:31.027 "num_base_bdevs_discovered": 1, 00:09:31.027 "num_base_bdevs_operational": 3, 00:09:31.027 "base_bdevs_list": [ 00:09:31.027 { 00:09:31.027 "name": null, 00:09:31.027 "uuid": "e83c6c66-9558-4d49-80fe-9da46fa19fa7", 00:09:31.027 "is_configured": false, 00:09:31.027 "data_offset": 0, 00:09:31.027 "data_size": 63488 00:09:31.027 }, 00:09:31.027 { 00:09:31.027 "name": null, 00:09:31.027 "uuid": "584debc2-8e92-4a6d-bfb2-005bcb6715fb", 00:09:31.027 "is_configured": false, 00:09:31.027 "data_offset": 0, 00:09:31.027 "data_size": 63488 00:09:31.027 }, 00:09:31.027 { 00:09:31.027 "name": "BaseBdev3", 00:09:31.027 "uuid": "1eacf62c-2995-4885-b370-0c7be5eb9f68", 00:09:31.027 "is_configured": true, 00:09:31.027 "data_offset": 2048, 00:09:31.027 "data_size": 63488 00:09:31.027 } 00:09:31.027 ] 00:09:31.027 }' 00:09:31.027 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:31.027 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:31.287 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:31.287 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:31.287 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:31.287 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:31.287 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:31.287 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:09:31.287 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:09:31.287 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:31.287 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:31.287 [2024-11-19 15:15:21.540483] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:31.287 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:31.287 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:09:31.287 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:31.287 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:31.287 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:31.287 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:31.287 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:31.287 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:31.287 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:31.287 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:31.287 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:31.287 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:31.287 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:31.287 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:31.287 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:31.287 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:31.287 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:31.287 "name": "Existed_Raid", 00:09:31.287 "uuid": "6c39dcfc-691f-4ca5-ad3f-a1438f4b24e7", 00:09:31.287 "strip_size_kb": 64, 00:09:31.287 "state": "configuring", 00:09:31.287 "raid_level": "raid0", 00:09:31.287 "superblock": true, 00:09:31.287 "num_base_bdevs": 3, 00:09:31.287 "num_base_bdevs_discovered": 2, 00:09:31.287 "num_base_bdevs_operational": 3, 00:09:31.287 "base_bdevs_list": [ 00:09:31.287 { 00:09:31.287 "name": null, 00:09:31.287 "uuid": "e83c6c66-9558-4d49-80fe-9da46fa19fa7", 00:09:31.287 "is_configured": false, 00:09:31.287 "data_offset": 0, 00:09:31.287 "data_size": 63488 00:09:31.287 }, 00:09:31.287 { 00:09:31.287 "name": "BaseBdev2", 00:09:31.287 "uuid": "584debc2-8e92-4a6d-bfb2-005bcb6715fb", 00:09:31.287 "is_configured": true, 00:09:31.287 "data_offset": 2048, 00:09:31.287 "data_size": 63488 00:09:31.287 }, 00:09:31.287 { 00:09:31.287 "name": "BaseBdev3", 00:09:31.287 "uuid": "1eacf62c-2995-4885-b370-0c7be5eb9f68", 00:09:31.287 "is_configured": true, 00:09:31.287 "data_offset": 2048, 00:09:31.287 "data_size": 63488 00:09:31.287 } 00:09:31.287 ] 00:09:31.287 }' 00:09:31.287 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:31.287 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:31.856 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:31.856 15:15:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:31.856 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:31.856 15:15:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:31.856 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:31.856 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:09:31.856 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:31.856 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:09:31.856 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:31.856 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:31.856 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:31.856 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u e83c6c66-9558-4d49-80fe-9da46fa19fa7 00:09:31.856 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:31.856 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:31.856 [2024-11-19 15:15:22.104531] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:09:31.856 [2024-11-19 15:15:22.104739] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:09:31.856 [2024-11-19 15:15:22.104783] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:09:31.856 NewBaseBdev 00:09:31.856 [2024-11-19 15:15:22.105083] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:09:31.856 [2024-11-19 15:15:22.105223] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:09:31.856 [2024-11-19 15:15:22.105233] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:09:31.856 [2024-11-19 15:15:22.105361] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:31.856 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:31.856 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:09:31.856 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:09:31.856 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:31.856 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:31.856 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:31.856 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:31.856 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:31.856 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:31.856 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:31.856 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:31.856 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:09:31.856 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:31.856 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:31.856 [ 00:09:31.856 { 00:09:31.856 "name": "NewBaseBdev", 00:09:31.856 "aliases": [ 00:09:31.856 "e83c6c66-9558-4d49-80fe-9da46fa19fa7" 00:09:31.856 ], 00:09:31.856 "product_name": "Malloc disk", 00:09:31.856 "block_size": 512, 00:09:31.856 "num_blocks": 65536, 00:09:31.856 "uuid": "e83c6c66-9558-4d49-80fe-9da46fa19fa7", 00:09:31.856 "assigned_rate_limits": { 00:09:31.856 "rw_ios_per_sec": 0, 00:09:31.856 "rw_mbytes_per_sec": 0, 00:09:31.856 "r_mbytes_per_sec": 0, 00:09:31.856 "w_mbytes_per_sec": 0 00:09:31.856 }, 00:09:31.856 "claimed": true, 00:09:31.856 "claim_type": "exclusive_write", 00:09:31.856 "zoned": false, 00:09:31.856 "supported_io_types": { 00:09:31.857 "read": true, 00:09:31.857 "write": true, 00:09:31.857 "unmap": true, 00:09:31.857 "flush": true, 00:09:31.857 "reset": true, 00:09:31.857 "nvme_admin": false, 00:09:31.857 "nvme_io": false, 00:09:31.857 "nvme_io_md": false, 00:09:31.857 "write_zeroes": true, 00:09:31.857 "zcopy": true, 00:09:31.857 "get_zone_info": false, 00:09:31.857 "zone_management": false, 00:09:31.857 "zone_append": false, 00:09:31.857 "compare": false, 00:09:31.857 "compare_and_write": false, 00:09:31.857 "abort": true, 00:09:31.857 "seek_hole": false, 00:09:31.857 "seek_data": false, 00:09:31.857 "copy": true, 00:09:31.857 "nvme_iov_md": false 00:09:31.857 }, 00:09:31.857 "memory_domains": [ 00:09:31.857 { 00:09:31.857 "dma_device_id": "system", 00:09:31.857 "dma_device_type": 1 00:09:31.857 }, 00:09:31.857 { 00:09:31.857 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:31.857 "dma_device_type": 2 00:09:31.857 } 00:09:31.857 ], 00:09:31.857 "driver_specific": {} 00:09:31.857 } 00:09:31.857 ] 00:09:31.857 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:31.857 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:31.857 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:09:31.857 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:31.857 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:31.857 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:31.857 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:31.857 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:31.857 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:31.857 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:31.857 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:31.857 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:31.857 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:31.857 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:31.857 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:31.857 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:31.857 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:32.127 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:32.127 "name": "Existed_Raid", 00:09:32.127 "uuid": "6c39dcfc-691f-4ca5-ad3f-a1438f4b24e7", 00:09:32.127 "strip_size_kb": 64, 00:09:32.127 "state": "online", 00:09:32.127 "raid_level": "raid0", 00:09:32.127 "superblock": true, 00:09:32.127 "num_base_bdevs": 3, 00:09:32.127 "num_base_bdevs_discovered": 3, 00:09:32.127 "num_base_bdevs_operational": 3, 00:09:32.127 "base_bdevs_list": [ 00:09:32.127 { 00:09:32.127 "name": "NewBaseBdev", 00:09:32.127 "uuid": "e83c6c66-9558-4d49-80fe-9da46fa19fa7", 00:09:32.127 "is_configured": true, 00:09:32.127 "data_offset": 2048, 00:09:32.127 "data_size": 63488 00:09:32.127 }, 00:09:32.127 { 00:09:32.127 "name": "BaseBdev2", 00:09:32.127 "uuid": "584debc2-8e92-4a6d-bfb2-005bcb6715fb", 00:09:32.127 "is_configured": true, 00:09:32.127 "data_offset": 2048, 00:09:32.127 "data_size": 63488 00:09:32.127 }, 00:09:32.127 { 00:09:32.127 "name": "BaseBdev3", 00:09:32.127 "uuid": "1eacf62c-2995-4885-b370-0c7be5eb9f68", 00:09:32.127 "is_configured": true, 00:09:32.127 "data_offset": 2048, 00:09:32.127 "data_size": 63488 00:09:32.127 } 00:09:32.127 ] 00:09:32.127 }' 00:09:32.127 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:32.127 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.401 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:09:32.401 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:32.401 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:32.401 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:32.401 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:09:32.401 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:32.401 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:32.401 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:32.401 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:32.401 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.401 [2024-11-19 15:15:22.584132] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:32.401 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:32.401 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:32.401 "name": "Existed_Raid", 00:09:32.401 "aliases": [ 00:09:32.401 "6c39dcfc-691f-4ca5-ad3f-a1438f4b24e7" 00:09:32.401 ], 00:09:32.401 "product_name": "Raid Volume", 00:09:32.401 "block_size": 512, 00:09:32.401 "num_blocks": 190464, 00:09:32.401 "uuid": "6c39dcfc-691f-4ca5-ad3f-a1438f4b24e7", 00:09:32.401 "assigned_rate_limits": { 00:09:32.401 "rw_ios_per_sec": 0, 00:09:32.401 "rw_mbytes_per_sec": 0, 00:09:32.401 "r_mbytes_per_sec": 0, 00:09:32.401 "w_mbytes_per_sec": 0 00:09:32.401 }, 00:09:32.401 "claimed": false, 00:09:32.401 "zoned": false, 00:09:32.401 "supported_io_types": { 00:09:32.401 "read": true, 00:09:32.401 "write": true, 00:09:32.401 "unmap": true, 00:09:32.401 "flush": true, 00:09:32.401 "reset": true, 00:09:32.401 "nvme_admin": false, 00:09:32.401 "nvme_io": false, 00:09:32.401 "nvme_io_md": false, 00:09:32.401 "write_zeroes": true, 00:09:32.401 "zcopy": false, 00:09:32.401 "get_zone_info": false, 00:09:32.401 "zone_management": false, 00:09:32.401 "zone_append": false, 00:09:32.401 "compare": false, 00:09:32.401 "compare_and_write": false, 00:09:32.401 "abort": false, 00:09:32.401 "seek_hole": false, 00:09:32.401 "seek_data": false, 00:09:32.401 "copy": false, 00:09:32.401 "nvme_iov_md": false 00:09:32.401 }, 00:09:32.401 "memory_domains": [ 00:09:32.401 { 00:09:32.401 "dma_device_id": "system", 00:09:32.401 "dma_device_type": 1 00:09:32.401 }, 00:09:32.401 { 00:09:32.401 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:32.402 "dma_device_type": 2 00:09:32.402 }, 00:09:32.402 { 00:09:32.402 "dma_device_id": "system", 00:09:32.402 "dma_device_type": 1 00:09:32.402 }, 00:09:32.402 { 00:09:32.402 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:32.402 "dma_device_type": 2 00:09:32.402 }, 00:09:32.402 { 00:09:32.402 "dma_device_id": "system", 00:09:32.402 "dma_device_type": 1 00:09:32.402 }, 00:09:32.402 { 00:09:32.402 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:32.402 "dma_device_type": 2 00:09:32.402 } 00:09:32.402 ], 00:09:32.402 "driver_specific": { 00:09:32.402 "raid": { 00:09:32.402 "uuid": "6c39dcfc-691f-4ca5-ad3f-a1438f4b24e7", 00:09:32.402 "strip_size_kb": 64, 00:09:32.402 "state": "online", 00:09:32.402 "raid_level": "raid0", 00:09:32.402 "superblock": true, 00:09:32.402 "num_base_bdevs": 3, 00:09:32.402 "num_base_bdevs_discovered": 3, 00:09:32.402 "num_base_bdevs_operational": 3, 00:09:32.402 "base_bdevs_list": [ 00:09:32.402 { 00:09:32.402 "name": "NewBaseBdev", 00:09:32.402 "uuid": "e83c6c66-9558-4d49-80fe-9da46fa19fa7", 00:09:32.402 "is_configured": true, 00:09:32.402 "data_offset": 2048, 00:09:32.402 "data_size": 63488 00:09:32.402 }, 00:09:32.402 { 00:09:32.402 "name": "BaseBdev2", 00:09:32.402 "uuid": "584debc2-8e92-4a6d-bfb2-005bcb6715fb", 00:09:32.402 "is_configured": true, 00:09:32.402 "data_offset": 2048, 00:09:32.402 "data_size": 63488 00:09:32.402 }, 00:09:32.402 { 00:09:32.402 "name": "BaseBdev3", 00:09:32.402 "uuid": "1eacf62c-2995-4885-b370-0c7be5eb9f68", 00:09:32.402 "is_configured": true, 00:09:32.402 "data_offset": 2048, 00:09:32.402 "data_size": 63488 00:09:32.402 } 00:09:32.402 ] 00:09:32.402 } 00:09:32.402 } 00:09:32.402 }' 00:09:32.402 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:32.402 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:09:32.402 BaseBdev2 00:09:32.402 BaseBdev3' 00:09:32.402 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:32.402 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:32.402 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:32.402 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:32.402 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:09:32.402 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:32.402 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.402 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:32.402 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:32.402 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:32.402 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:32.402 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:32.402 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:32.402 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.402 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:32.402 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:32.662 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:32.662 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:32.662 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:32.662 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:32.662 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:32.662 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.662 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:32.662 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:32.662 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:32.662 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:32.662 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:32.662 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:32.662 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.662 [2024-11-19 15:15:22.811345] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:32.662 [2024-11-19 15:15:22.811382] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:32.662 [2024-11-19 15:15:22.811472] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:32.662 [2024-11-19 15:15:22.811535] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:32.662 [2024-11-19 15:15:22.811554] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:09:32.662 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:32.662 15:15:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 75629 00:09:32.662 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 75629 ']' 00:09:32.662 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 75629 00:09:32.662 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:09:32.662 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:32.662 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 75629 00:09:32.662 killing process with pid 75629 00:09:32.662 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:32.662 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:32.662 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 75629' 00:09:32.662 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 75629 00:09:32.662 [2024-11-19 15:15:22.862671] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:32.662 15:15:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 75629 00:09:32.662 [2024-11-19 15:15:22.919836] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:32.922 ************************************ 00:09:32.922 END TEST raid_state_function_test_sb 00:09:32.922 15:15:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:09:32.922 00:09:32.922 real 0m8.920s 00:09:32.922 user 0m14.992s 00:09:32.922 sys 0m1.807s 00:09:32.922 15:15:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:32.922 15:15:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.922 ************************************ 00:09:33.181 15:15:23 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid0 3 00:09:33.181 15:15:23 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:09:33.181 15:15:23 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:33.181 15:15:23 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:33.181 ************************************ 00:09:33.181 START TEST raid_superblock_test 00:09:33.181 ************************************ 00:09:33.181 15:15:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid0 3 00:09:33.181 15:15:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid0 00:09:33.182 15:15:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:09:33.182 15:15:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:09:33.182 15:15:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:09:33.182 15:15:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:09:33.182 15:15:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:09:33.182 15:15:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:09:33.182 15:15:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:09:33.182 15:15:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:09:33.182 15:15:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:09:33.182 15:15:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:09:33.182 15:15:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:09:33.182 15:15:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:09:33.182 15:15:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid0 '!=' raid1 ']' 00:09:33.182 15:15:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:09:33.182 15:15:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:09:33.182 15:15:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=76233 00:09:33.182 15:15:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:09:33.182 15:15:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 76233 00:09:33.182 15:15:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 76233 ']' 00:09:33.182 15:15:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:33.182 15:15:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:33.182 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:33.182 15:15:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:33.182 15:15:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:33.182 15:15:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:33.182 [2024-11-19 15:15:23.405847] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:09:33.182 [2024-11-19 15:15:23.406022] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid76233 ] 00:09:33.441 [2024-11-19 15:15:23.558989] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:33.441 [2024-11-19 15:15:23.597275] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:33.441 [2024-11-19 15:15:23.673561] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:33.441 [2024-11-19 15:15:23.673608] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.011 malloc1 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.011 [2024-11-19 15:15:24.271390] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:34.011 [2024-11-19 15:15:24.271465] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:34.011 [2024-11-19 15:15:24.271487] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:09:34.011 [2024-11-19 15:15:24.271503] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:34.011 [2024-11-19 15:15:24.274066] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:34.011 [2024-11-19 15:15:24.274109] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:34.011 pt1 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.011 malloc2 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.011 [2024-11-19 15:15:24.306004] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:34.011 [2024-11-19 15:15:24.306054] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:34.011 [2024-11-19 15:15:24.306070] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:09:34.011 [2024-11-19 15:15:24.306080] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:34.011 [2024-11-19 15:15:24.308492] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:34.011 [2024-11-19 15:15:24.308526] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:34.011 pt2 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.011 malloc3 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.011 15:15:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.012 [2024-11-19 15:15:24.340500] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:34.012 [2024-11-19 15:15:24.340562] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:34.012 [2024-11-19 15:15:24.340582] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:09:34.012 [2024-11-19 15:15:24.340594] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:34.012 [2024-11-19 15:15:24.343038] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:34.012 [2024-11-19 15:15:24.343073] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:34.012 pt3 00:09:34.012 15:15:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.012 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:34.012 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:34.012 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:09:34.012 15:15:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.012 15:15:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.272 [2024-11-19 15:15:24.352557] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:34.272 [2024-11-19 15:15:24.354777] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:34.272 [2024-11-19 15:15:24.354854] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:34.272 [2024-11-19 15:15:24.355048] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:09:34.272 [2024-11-19 15:15:24.355061] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:09:34.272 [2024-11-19 15:15:24.355366] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:09:34.272 [2024-11-19 15:15:24.355522] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:09:34.272 [2024-11-19 15:15:24.355542] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:09:34.272 [2024-11-19 15:15:24.355686] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:34.272 15:15:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.272 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:09:34.272 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:34.272 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:34.272 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:34.272 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:34.272 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:34.272 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:34.272 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:34.272 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:34.272 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:34.272 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:34.272 15:15:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.272 15:15:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.272 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:34.272 15:15:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.272 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:34.272 "name": "raid_bdev1", 00:09:34.272 "uuid": "f9971316-f7ef-42b1-848a-1e06a62dd414", 00:09:34.272 "strip_size_kb": 64, 00:09:34.272 "state": "online", 00:09:34.272 "raid_level": "raid0", 00:09:34.272 "superblock": true, 00:09:34.272 "num_base_bdevs": 3, 00:09:34.272 "num_base_bdevs_discovered": 3, 00:09:34.272 "num_base_bdevs_operational": 3, 00:09:34.272 "base_bdevs_list": [ 00:09:34.272 { 00:09:34.272 "name": "pt1", 00:09:34.272 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:34.272 "is_configured": true, 00:09:34.272 "data_offset": 2048, 00:09:34.272 "data_size": 63488 00:09:34.272 }, 00:09:34.272 { 00:09:34.272 "name": "pt2", 00:09:34.272 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:34.272 "is_configured": true, 00:09:34.272 "data_offset": 2048, 00:09:34.272 "data_size": 63488 00:09:34.272 }, 00:09:34.272 { 00:09:34.272 "name": "pt3", 00:09:34.272 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:34.272 "is_configured": true, 00:09:34.272 "data_offset": 2048, 00:09:34.272 "data_size": 63488 00:09:34.272 } 00:09:34.272 ] 00:09:34.272 }' 00:09:34.272 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:34.272 15:15:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.532 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:09:34.532 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:09:34.532 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:34.532 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:34.532 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:34.532 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:34.532 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:34.532 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:34.532 15:15:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.532 15:15:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.532 [2024-11-19 15:15:24.860189] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:34.903 15:15:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.903 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:34.903 "name": "raid_bdev1", 00:09:34.903 "aliases": [ 00:09:34.903 "f9971316-f7ef-42b1-848a-1e06a62dd414" 00:09:34.903 ], 00:09:34.903 "product_name": "Raid Volume", 00:09:34.903 "block_size": 512, 00:09:34.903 "num_blocks": 190464, 00:09:34.903 "uuid": "f9971316-f7ef-42b1-848a-1e06a62dd414", 00:09:34.903 "assigned_rate_limits": { 00:09:34.903 "rw_ios_per_sec": 0, 00:09:34.903 "rw_mbytes_per_sec": 0, 00:09:34.903 "r_mbytes_per_sec": 0, 00:09:34.903 "w_mbytes_per_sec": 0 00:09:34.903 }, 00:09:34.903 "claimed": false, 00:09:34.903 "zoned": false, 00:09:34.903 "supported_io_types": { 00:09:34.903 "read": true, 00:09:34.903 "write": true, 00:09:34.903 "unmap": true, 00:09:34.903 "flush": true, 00:09:34.903 "reset": true, 00:09:34.903 "nvme_admin": false, 00:09:34.903 "nvme_io": false, 00:09:34.903 "nvme_io_md": false, 00:09:34.903 "write_zeroes": true, 00:09:34.903 "zcopy": false, 00:09:34.903 "get_zone_info": false, 00:09:34.903 "zone_management": false, 00:09:34.903 "zone_append": false, 00:09:34.903 "compare": false, 00:09:34.903 "compare_and_write": false, 00:09:34.903 "abort": false, 00:09:34.903 "seek_hole": false, 00:09:34.903 "seek_data": false, 00:09:34.903 "copy": false, 00:09:34.903 "nvme_iov_md": false 00:09:34.903 }, 00:09:34.903 "memory_domains": [ 00:09:34.903 { 00:09:34.903 "dma_device_id": "system", 00:09:34.903 "dma_device_type": 1 00:09:34.903 }, 00:09:34.903 { 00:09:34.903 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:34.903 "dma_device_type": 2 00:09:34.903 }, 00:09:34.903 { 00:09:34.903 "dma_device_id": "system", 00:09:34.903 "dma_device_type": 1 00:09:34.903 }, 00:09:34.903 { 00:09:34.903 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:34.903 "dma_device_type": 2 00:09:34.903 }, 00:09:34.903 { 00:09:34.903 "dma_device_id": "system", 00:09:34.903 "dma_device_type": 1 00:09:34.903 }, 00:09:34.903 { 00:09:34.903 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:34.903 "dma_device_type": 2 00:09:34.903 } 00:09:34.903 ], 00:09:34.903 "driver_specific": { 00:09:34.903 "raid": { 00:09:34.903 "uuid": "f9971316-f7ef-42b1-848a-1e06a62dd414", 00:09:34.903 "strip_size_kb": 64, 00:09:34.903 "state": "online", 00:09:34.903 "raid_level": "raid0", 00:09:34.903 "superblock": true, 00:09:34.903 "num_base_bdevs": 3, 00:09:34.903 "num_base_bdevs_discovered": 3, 00:09:34.903 "num_base_bdevs_operational": 3, 00:09:34.903 "base_bdevs_list": [ 00:09:34.903 { 00:09:34.903 "name": "pt1", 00:09:34.903 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:34.903 "is_configured": true, 00:09:34.903 "data_offset": 2048, 00:09:34.903 "data_size": 63488 00:09:34.903 }, 00:09:34.903 { 00:09:34.903 "name": "pt2", 00:09:34.903 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:34.903 "is_configured": true, 00:09:34.903 "data_offset": 2048, 00:09:34.903 "data_size": 63488 00:09:34.903 }, 00:09:34.903 { 00:09:34.903 "name": "pt3", 00:09:34.903 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:34.903 "is_configured": true, 00:09:34.903 "data_offset": 2048, 00:09:34.903 "data_size": 63488 00:09:34.903 } 00:09:34.903 ] 00:09:34.903 } 00:09:34.903 } 00:09:34.903 }' 00:09:34.903 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:34.903 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:09:34.903 pt2 00:09:34.903 pt3' 00:09:34.903 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:34.903 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:34.903 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:34.903 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:09:34.903 15:15:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.903 15:15:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:34.903 15:15:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.903 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.903 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:34.903 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:34.903 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:34.903 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:09:34.903 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.903 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.903 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:34.903 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.903 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:34.903 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:34.903 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:34.903 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:09:34.903 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.903 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.903 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:34.903 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.903 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:34.903 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:34.903 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:09:34.903 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:34.903 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.904 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.904 [2024-11-19 15:15:25.135563] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:34.904 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.904 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=f9971316-f7ef-42b1-848a-1e06a62dd414 00:09:34.904 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z f9971316-f7ef-42b1-848a-1e06a62dd414 ']' 00:09:34.904 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:34.904 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.904 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.904 [2024-11-19 15:15:25.175265] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:34.904 [2024-11-19 15:15:25.175305] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:34.904 [2024-11-19 15:15:25.175413] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:34.904 [2024-11-19 15:15:25.175494] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:34.904 [2024-11-19 15:15:25.175511] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:09:34.904 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.904 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:34.904 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:09:34.904 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.904 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.904 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.904 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:09:34.904 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:09:34.904 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:34.904 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:09:34.904 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.904 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.904 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.172 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:35.172 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:09:35.172 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.172 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.172 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.172 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:35.172 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:09:35.172 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.172 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.172 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.172 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:09:35.172 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:09:35.172 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.172 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.172 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.172 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:09:35.172 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:09:35.172 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:09:35.172 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:09:35.172 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:09:35.172 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:09:35.172 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:09:35.172 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:09:35.172 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:09:35.172 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.172 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.172 [2024-11-19 15:15:25.327070] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:09:35.172 [2024-11-19 15:15:25.329382] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:09:35.172 [2024-11-19 15:15:25.329492] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:09:35.172 [2024-11-19 15:15:25.329569] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:09:35.172 [2024-11-19 15:15:25.329654] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:09:35.172 [2024-11-19 15:15:25.329723] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:09:35.172 [2024-11-19 15:15:25.329739] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:35.172 [2024-11-19 15:15:25.329751] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:09:35.172 request: 00:09:35.172 { 00:09:35.172 "name": "raid_bdev1", 00:09:35.172 "raid_level": "raid0", 00:09:35.172 "base_bdevs": [ 00:09:35.172 "malloc1", 00:09:35.172 "malloc2", 00:09:35.172 "malloc3" 00:09:35.172 ], 00:09:35.172 "strip_size_kb": 64, 00:09:35.172 "superblock": false, 00:09:35.172 "method": "bdev_raid_create", 00:09:35.172 "req_id": 1 00:09:35.172 } 00:09:35.172 Got JSON-RPC error response 00:09:35.172 response: 00:09:35.172 { 00:09:35.172 "code": -17, 00:09:35.172 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:09:35.172 } 00:09:35.172 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:09:35.172 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:09:35.172 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:09:35.172 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:09:35.172 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:09:35.172 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:35.172 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:09:35.173 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.173 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.173 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.173 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:09:35.173 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:09:35.173 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:35.173 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.173 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.173 [2024-11-19 15:15:25.390905] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:35.173 [2024-11-19 15:15:25.391001] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:35.173 [2024-11-19 15:15:25.391037] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:09:35.173 [2024-11-19 15:15:25.391071] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:35.173 [2024-11-19 15:15:25.393580] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:35.173 [2024-11-19 15:15:25.393655] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:35.173 [2024-11-19 15:15:25.393740] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:09:35.173 [2024-11-19 15:15:25.393801] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:35.173 pt1 00:09:35.173 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.173 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 3 00:09:35.173 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:35.173 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:35.173 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:35.173 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:35.173 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:35.173 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:35.173 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:35.173 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:35.173 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:35.173 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:35.173 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:35.173 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.173 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.173 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.173 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:35.173 "name": "raid_bdev1", 00:09:35.173 "uuid": "f9971316-f7ef-42b1-848a-1e06a62dd414", 00:09:35.173 "strip_size_kb": 64, 00:09:35.173 "state": "configuring", 00:09:35.173 "raid_level": "raid0", 00:09:35.173 "superblock": true, 00:09:35.173 "num_base_bdevs": 3, 00:09:35.173 "num_base_bdevs_discovered": 1, 00:09:35.173 "num_base_bdevs_operational": 3, 00:09:35.173 "base_bdevs_list": [ 00:09:35.173 { 00:09:35.173 "name": "pt1", 00:09:35.173 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:35.173 "is_configured": true, 00:09:35.173 "data_offset": 2048, 00:09:35.173 "data_size": 63488 00:09:35.173 }, 00:09:35.173 { 00:09:35.173 "name": null, 00:09:35.173 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:35.173 "is_configured": false, 00:09:35.173 "data_offset": 2048, 00:09:35.173 "data_size": 63488 00:09:35.173 }, 00:09:35.173 { 00:09:35.173 "name": null, 00:09:35.173 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:35.173 "is_configured": false, 00:09:35.173 "data_offset": 2048, 00:09:35.173 "data_size": 63488 00:09:35.173 } 00:09:35.173 ] 00:09:35.173 }' 00:09:35.173 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:35.173 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.739 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:09:35.739 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:35.739 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.739 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.739 [2024-11-19 15:15:25.806264] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:35.739 [2024-11-19 15:15:25.806366] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:35.739 [2024-11-19 15:15:25.806393] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:09:35.739 [2024-11-19 15:15:25.806409] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:35.740 [2024-11-19 15:15:25.806893] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:35.740 [2024-11-19 15:15:25.806914] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:35.740 [2024-11-19 15:15:25.807027] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:35.740 [2024-11-19 15:15:25.807062] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:35.740 pt2 00:09:35.740 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.740 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:09:35.740 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.740 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.740 [2024-11-19 15:15:25.818225] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:09:35.740 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.740 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 3 00:09:35.740 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:35.740 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:35.740 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:35.740 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:35.740 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:35.740 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:35.740 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:35.740 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:35.740 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:35.740 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:35.740 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:35.740 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.740 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.740 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.740 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:35.740 "name": "raid_bdev1", 00:09:35.740 "uuid": "f9971316-f7ef-42b1-848a-1e06a62dd414", 00:09:35.740 "strip_size_kb": 64, 00:09:35.740 "state": "configuring", 00:09:35.740 "raid_level": "raid0", 00:09:35.740 "superblock": true, 00:09:35.740 "num_base_bdevs": 3, 00:09:35.740 "num_base_bdevs_discovered": 1, 00:09:35.740 "num_base_bdevs_operational": 3, 00:09:35.740 "base_bdevs_list": [ 00:09:35.740 { 00:09:35.740 "name": "pt1", 00:09:35.740 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:35.740 "is_configured": true, 00:09:35.740 "data_offset": 2048, 00:09:35.740 "data_size": 63488 00:09:35.740 }, 00:09:35.740 { 00:09:35.740 "name": null, 00:09:35.740 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:35.740 "is_configured": false, 00:09:35.740 "data_offset": 0, 00:09:35.740 "data_size": 63488 00:09:35.740 }, 00:09:35.740 { 00:09:35.740 "name": null, 00:09:35.740 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:35.740 "is_configured": false, 00:09:35.740 "data_offset": 2048, 00:09:35.740 "data_size": 63488 00:09:35.740 } 00:09:35.740 ] 00:09:35.740 }' 00:09:35.740 15:15:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:35.740 15:15:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.998 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:09:35.998 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:35.998 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:35.998 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.998 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.998 [2024-11-19 15:15:26.233536] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:35.998 [2024-11-19 15:15:26.233680] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:35.998 [2024-11-19 15:15:26.233725] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:09:35.998 [2024-11-19 15:15:26.233758] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:35.998 [2024-11-19 15:15:26.234322] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:35.998 [2024-11-19 15:15:26.234385] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:35.998 [2024-11-19 15:15:26.234512] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:35.998 [2024-11-19 15:15:26.234567] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:35.998 pt2 00:09:35.998 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.998 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:09:35.998 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:35.998 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:35.998 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.998 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.998 [2024-11-19 15:15:26.245465] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:35.998 [2024-11-19 15:15:26.245548] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:35.998 [2024-11-19 15:15:26.245585] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:09:35.998 [2024-11-19 15:15:26.245611] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:35.998 [2024-11-19 15:15:26.246034] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:35.998 [2024-11-19 15:15:26.246089] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:35.999 [2024-11-19 15:15:26.246192] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:09:35.999 [2024-11-19 15:15:26.246236] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:35.999 [2024-11-19 15:15:26.246354] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:09:35.999 [2024-11-19 15:15:26.246389] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:09:35.999 [2024-11-19 15:15:26.246662] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:09:35.999 [2024-11-19 15:15:26.246809] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:09:35.999 [2024-11-19 15:15:26.246850] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:09:35.999 [2024-11-19 15:15:26.247008] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:35.999 pt3 00:09:35.999 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.999 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:09:35.999 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:35.999 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:09:35.999 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:35.999 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:35.999 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:35.999 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:35.999 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:35.999 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:35.999 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:35.999 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:35.999 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:35.999 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:35.999 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:35.999 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.999 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:35.999 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:35.999 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:35.999 "name": "raid_bdev1", 00:09:35.999 "uuid": "f9971316-f7ef-42b1-848a-1e06a62dd414", 00:09:35.999 "strip_size_kb": 64, 00:09:35.999 "state": "online", 00:09:35.999 "raid_level": "raid0", 00:09:35.999 "superblock": true, 00:09:35.999 "num_base_bdevs": 3, 00:09:35.999 "num_base_bdevs_discovered": 3, 00:09:35.999 "num_base_bdevs_operational": 3, 00:09:35.999 "base_bdevs_list": [ 00:09:35.999 { 00:09:35.999 "name": "pt1", 00:09:35.999 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:35.999 "is_configured": true, 00:09:35.999 "data_offset": 2048, 00:09:35.999 "data_size": 63488 00:09:35.999 }, 00:09:35.999 { 00:09:35.999 "name": "pt2", 00:09:35.999 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:35.999 "is_configured": true, 00:09:35.999 "data_offset": 2048, 00:09:35.999 "data_size": 63488 00:09:35.999 }, 00:09:35.999 { 00:09:35.999 "name": "pt3", 00:09:35.999 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:35.999 "is_configured": true, 00:09:35.999 "data_offset": 2048, 00:09:35.999 "data_size": 63488 00:09:35.999 } 00:09:35.999 ] 00:09:35.999 }' 00:09:35.999 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:35.999 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.566 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:09:36.566 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:09:36.566 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:36.566 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:36.566 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:36.566 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:36.566 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:36.566 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:36.566 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.566 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.566 [2024-11-19 15:15:26.733012] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:36.566 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.566 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:36.566 "name": "raid_bdev1", 00:09:36.566 "aliases": [ 00:09:36.566 "f9971316-f7ef-42b1-848a-1e06a62dd414" 00:09:36.566 ], 00:09:36.566 "product_name": "Raid Volume", 00:09:36.566 "block_size": 512, 00:09:36.566 "num_blocks": 190464, 00:09:36.566 "uuid": "f9971316-f7ef-42b1-848a-1e06a62dd414", 00:09:36.566 "assigned_rate_limits": { 00:09:36.566 "rw_ios_per_sec": 0, 00:09:36.566 "rw_mbytes_per_sec": 0, 00:09:36.566 "r_mbytes_per_sec": 0, 00:09:36.566 "w_mbytes_per_sec": 0 00:09:36.566 }, 00:09:36.566 "claimed": false, 00:09:36.566 "zoned": false, 00:09:36.566 "supported_io_types": { 00:09:36.566 "read": true, 00:09:36.566 "write": true, 00:09:36.566 "unmap": true, 00:09:36.566 "flush": true, 00:09:36.566 "reset": true, 00:09:36.566 "nvme_admin": false, 00:09:36.566 "nvme_io": false, 00:09:36.566 "nvme_io_md": false, 00:09:36.566 "write_zeroes": true, 00:09:36.566 "zcopy": false, 00:09:36.566 "get_zone_info": false, 00:09:36.566 "zone_management": false, 00:09:36.566 "zone_append": false, 00:09:36.566 "compare": false, 00:09:36.566 "compare_and_write": false, 00:09:36.566 "abort": false, 00:09:36.566 "seek_hole": false, 00:09:36.566 "seek_data": false, 00:09:36.566 "copy": false, 00:09:36.566 "nvme_iov_md": false 00:09:36.566 }, 00:09:36.566 "memory_domains": [ 00:09:36.566 { 00:09:36.566 "dma_device_id": "system", 00:09:36.566 "dma_device_type": 1 00:09:36.566 }, 00:09:36.566 { 00:09:36.566 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:36.566 "dma_device_type": 2 00:09:36.566 }, 00:09:36.566 { 00:09:36.566 "dma_device_id": "system", 00:09:36.566 "dma_device_type": 1 00:09:36.566 }, 00:09:36.566 { 00:09:36.566 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:36.566 "dma_device_type": 2 00:09:36.566 }, 00:09:36.566 { 00:09:36.566 "dma_device_id": "system", 00:09:36.566 "dma_device_type": 1 00:09:36.566 }, 00:09:36.566 { 00:09:36.566 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:36.566 "dma_device_type": 2 00:09:36.566 } 00:09:36.566 ], 00:09:36.566 "driver_specific": { 00:09:36.566 "raid": { 00:09:36.566 "uuid": "f9971316-f7ef-42b1-848a-1e06a62dd414", 00:09:36.566 "strip_size_kb": 64, 00:09:36.566 "state": "online", 00:09:36.566 "raid_level": "raid0", 00:09:36.566 "superblock": true, 00:09:36.566 "num_base_bdevs": 3, 00:09:36.566 "num_base_bdevs_discovered": 3, 00:09:36.566 "num_base_bdevs_operational": 3, 00:09:36.566 "base_bdevs_list": [ 00:09:36.566 { 00:09:36.566 "name": "pt1", 00:09:36.566 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:36.566 "is_configured": true, 00:09:36.566 "data_offset": 2048, 00:09:36.566 "data_size": 63488 00:09:36.566 }, 00:09:36.566 { 00:09:36.566 "name": "pt2", 00:09:36.566 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:36.566 "is_configured": true, 00:09:36.566 "data_offset": 2048, 00:09:36.566 "data_size": 63488 00:09:36.566 }, 00:09:36.566 { 00:09:36.566 "name": "pt3", 00:09:36.566 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:36.566 "is_configured": true, 00:09:36.566 "data_offset": 2048, 00:09:36.566 "data_size": 63488 00:09:36.566 } 00:09:36.566 ] 00:09:36.566 } 00:09:36.566 } 00:09:36.566 }' 00:09:36.566 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:36.566 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:09:36.566 pt2 00:09:36.566 pt3' 00:09:36.566 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:36.566 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:36.566 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:36.566 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:09:36.566 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:36.566 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.566 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.566 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.566 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:36.566 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:36.566 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:36.566 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:09:36.566 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:36.566 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.566 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.826 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.826 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:36.826 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:36.826 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:36.826 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:09:36.826 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:36.826 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.826 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.827 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.827 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:36.827 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:36.827 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:09:36.827 15:15:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:36.827 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.827 15:15:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.827 [2024-11-19 15:15:27.000451] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:36.827 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.827 15:15:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' f9971316-f7ef-42b1-848a-1e06a62dd414 '!=' f9971316-f7ef-42b1-848a-1e06a62dd414 ']' 00:09:36.827 15:15:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid0 00:09:36.827 15:15:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:36.827 15:15:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:36.827 15:15:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 76233 00:09:36.827 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 76233 ']' 00:09:36.827 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 76233 00:09:36.827 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:09:36.827 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:36.827 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 76233 00:09:36.827 killing process with pid 76233 00:09:36.827 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:36.827 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:36.827 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 76233' 00:09:36.827 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 76233 00:09:36.827 [2024-11-19 15:15:27.069243] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:36.827 [2024-11-19 15:15:27.069334] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:36.827 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 76233 00:09:36.827 [2024-11-19 15:15:27.069405] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:36.827 [2024-11-19 15:15:27.069414] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:09:36.827 [2024-11-19 15:15:27.129120] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:37.396 15:15:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:09:37.396 00:09:37.396 real 0m4.132s 00:09:37.396 user 0m6.398s 00:09:37.396 sys 0m0.935s 00:09:37.396 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:37.396 15:15:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:37.396 ************************************ 00:09:37.396 END TEST raid_superblock_test 00:09:37.396 ************************************ 00:09:37.396 15:15:27 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid0 3 read 00:09:37.396 15:15:27 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:37.396 15:15:27 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:37.396 15:15:27 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:37.396 ************************************ 00:09:37.396 START TEST raid_read_error_test 00:09:37.396 ************************************ 00:09:37.396 15:15:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid0 3 read 00:09:37.396 15:15:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:09:37.396 15:15:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:09:37.396 15:15:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:09:37.396 15:15:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:09:37.396 15:15:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:37.396 15:15:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:09:37.396 15:15:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:37.396 15:15:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:37.396 15:15:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:09:37.396 15:15:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:37.396 15:15:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:37.396 15:15:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:09:37.396 15:15:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:37.396 15:15:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:37.396 15:15:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:09:37.396 15:15:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:09:37.396 15:15:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:09:37.396 15:15:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:09:37.396 15:15:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:09:37.396 15:15:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:09:37.396 15:15:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:09:37.396 15:15:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:09:37.396 15:15:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:09:37.396 15:15:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:09:37.396 15:15:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:09:37.396 15:15:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.4JsUMAgajk 00:09:37.396 15:15:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=76475 00:09:37.396 15:15:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:09:37.396 15:15:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 76475 00:09:37.396 15:15:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 76475 ']' 00:09:37.396 15:15:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:37.396 15:15:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:37.396 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:37.397 15:15:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:37.397 15:15:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:37.397 15:15:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:37.397 [2024-11-19 15:15:27.621359] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:09:37.397 [2024-11-19 15:15:27.621499] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid76475 ] 00:09:37.655 [2024-11-19 15:15:27.762072] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:37.655 [2024-11-19 15:15:27.802982] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:37.655 [2024-11-19 15:15:27.879102] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:37.655 [2024-11-19 15:15:27.879146] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:38.224 15:15:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:38.224 15:15:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:09:38.224 15:15:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:38.224 15:15:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:09:38.224 15:15:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.224 15:15:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.224 BaseBdev1_malloc 00:09:38.224 15:15:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.224 15:15:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:09:38.224 15:15:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.224 15:15:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.224 true 00:09:38.224 15:15:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.224 15:15:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:09:38.224 15:15:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.224 15:15:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.224 [2024-11-19 15:15:28.513581] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:09:38.224 [2024-11-19 15:15:28.513643] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:38.224 [2024-11-19 15:15:28.513679] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:09:38.224 [2024-11-19 15:15:28.513692] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:38.224 [2024-11-19 15:15:28.516259] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:38.224 [2024-11-19 15:15:28.516299] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:09:38.224 BaseBdev1 00:09:38.224 15:15:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.224 15:15:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:38.224 15:15:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:09:38.224 15:15:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.224 15:15:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.224 BaseBdev2_malloc 00:09:38.224 15:15:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.224 15:15:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:09:38.224 15:15:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.224 15:15:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.224 true 00:09:38.224 15:15:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.224 15:15:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:09:38.224 15:15:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.224 15:15:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.224 [2024-11-19 15:15:28.560391] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:09:38.224 [2024-11-19 15:15:28.560442] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:38.224 [2024-11-19 15:15:28.560461] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:09:38.224 [2024-11-19 15:15:28.560480] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:38.485 [2024-11-19 15:15:28.563081] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:38.485 [2024-11-19 15:15:28.563119] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:09:38.485 BaseBdev2 00:09:38.485 15:15:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.485 15:15:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:38.485 15:15:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:09:38.485 15:15:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.485 15:15:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.485 BaseBdev3_malloc 00:09:38.485 15:15:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.485 15:15:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:09:38.485 15:15:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.485 15:15:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.485 true 00:09:38.485 15:15:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.485 15:15:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:09:38.485 15:15:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.485 15:15:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.485 [2024-11-19 15:15:28.606999] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:09:38.485 [2024-11-19 15:15:28.607064] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:38.485 [2024-11-19 15:15:28.607084] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:09:38.485 [2024-11-19 15:15:28.607093] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:38.485 [2024-11-19 15:15:28.609540] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:38.485 [2024-11-19 15:15:28.609574] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:09:38.485 BaseBdev3 00:09:38.485 15:15:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.485 15:15:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:09:38.485 15:15:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.485 15:15:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.485 [2024-11-19 15:15:28.619074] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:38.485 [2024-11-19 15:15:28.621284] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:38.485 [2024-11-19 15:15:28.621376] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:38.485 [2024-11-19 15:15:28.621558] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:09:38.485 [2024-11-19 15:15:28.621573] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:09:38.485 [2024-11-19 15:15:28.621857] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002bb0 00:09:38.485 [2024-11-19 15:15:28.622050] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:09:38.485 [2024-11-19 15:15:28.622065] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:09:38.485 [2024-11-19 15:15:28.622209] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:38.485 15:15:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.485 15:15:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:09:38.485 15:15:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:38.485 15:15:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:38.485 15:15:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:38.485 15:15:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:38.485 15:15:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:38.485 15:15:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:38.485 15:15:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:38.485 15:15:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:38.485 15:15:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:38.485 15:15:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:38.485 15:15:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:38.485 15:15:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.485 15:15:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.485 15:15:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.485 15:15:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:38.485 "name": "raid_bdev1", 00:09:38.485 "uuid": "f1b491b4-bacb-4ef2-8bb6-9210d0ec5ef2", 00:09:38.485 "strip_size_kb": 64, 00:09:38.485 "state": "online", 00:09:38.485 "raid_level": "raid0", 00:09:38.485 "superblock": true, 00:09:38.485 "num_base_bdevs": 3, 00:09:38.485 "num_base_bdevs_discovered": 3, 00:09:38.485 "num_base_bdevs_operational": 3, 00:09:38.485 "base_bdevs_list": [ 00:09:38.485 { 00:09:38.485 "name": "BaseBdev1", 00:09:38.485 "uuid": "079fe795-de10-5bfd-945d-58595464895f", 00:09:38.485 "is_configured": true, 00:09:38.485 "data_offset": 2048, 00:09:38.485 "data_size": 63488 00:09:38.485 }, 00:09:38.485 { 00:09:38.485 "name": "BaseBdev2", 00:09:38.485 "uuid": "d17ef0de-b24e-5250-99b9-a43797402a60", 00:09:38.485 "is_configured": true, 00:09:38.485 "data_offset": 2048, 00:09:38.485 "data_size": 63488 00:09:38.485 }, 00:09:38.485 { 00:09:38.485 "name": "BaseBdev3", 00:09:38.485 "uuid": "d5b8c7fe-4437-552c-a865-10d74b4fccea", 00:09:38.485 "is_configured": true, 00:09:38.485 "data_offset": 2048, 00:09:38.485 "data_size": 63488 00:09:38.485 } 00:09:38.485 ] 00:09:38.485 }' 00:09:38.485 15:15:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:38.485 15:15:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.745 15:15:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:09:38.745 15:15:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:09:39.013 [2024-11-19 15:15:29.170642] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002d50 00:09:39.955 15:15:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:09:39.955 15:15:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:39.955 15:15:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:39.955 15:15:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:39.955 15:15:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:09:39.955 15:15:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:09:39.955 15:15:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:09:39.955 15:15:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:09:39.955 15:15:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:39.955 15:15:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:39.955 15:15:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:39.955 15:15:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:39.955 15:15:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:39.955 15:15:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:39.955 15:15:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:39.955 15:15:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:39.955 15:15:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:39.955 15:15:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:39.955 15:15:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:39.955 15:15:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:39.955 15:15:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:39.955 15:15:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:39.955 15:15:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:39.955 "name": "raid_bdev1", 00:09:39.955 "uuid": "f1b491b4-bacb-4ef2-8bb6-9210d0ec5ef2", 00:09:39.955 "strip_size_kb": 64, 00:09:39.955 "state": "online", 00:09:39.955 "raid_level": "raid0", 00:09:39.955 "superblock": true, 00:09:39.955 "num_base_bdevs": 3, 00:09:39.955 "num_base_bdevs_discovered": 3, 00:09:39.955 "num_base_bdevs_operational": 3, 00:09:39.955 "base_bdevs_list": [ 00:09:39.955 { 00:09:39.955 "name": "BaseBdev1", 00:09:39.955 "uuid": "079fe795-de10-5bfd-945d-58595464895f", 00:09:39.955 "is_configured": true, 00:09:39.955 "data_offset": 2048, 00:09:39.955 "data_size": 63488 00:09:39.955 }, 00:09:39.955 { 00:09:39.955 "name": "BaseBdev2", 00:09:39.955 "uuid": "d17ef0de-b24e-5250-99b9-a43797402a60", 00:09:39.955 "is_configured": true, 00:09:39.955 "data_offset": 2048, 00:09:39.955 "data_size": 63488 00:09:39.955 }, 00:09:39.955 { 00:09:39.955 "name": "BaseBdev3", 00:09:39.955 "uuid": "d5b8c7fe-4437-552c-a865-10d74b4fccea", 00:09:39.956 "is_configured": true, 00:09:39.956 "data_offset": 2048, 00:09:39.956 "data_size": 63488 00:09:39.956 } 00:09:39.956 ] 00:09:39.956 }' 00:09:39.956 15:15:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:39.956 15:15:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.523 15:15:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:40.523 15:15:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:40.523 15:15:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.523 [2024-11-19 15:15:30.567441] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:40.523 [2024-11-19 15:15:30.567497] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:40.523 [2024-11-19 15:15:30.570164] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:40.523 [2024-11-19 15:15:30.570236] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:40.523 [2024-11-19 15:15:30.570274] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:40.523 [2024-11-19 15:15:30.570293] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:09:40.523 { 00:09:40.523 "results": [ 00:09:40.523 { 00:09:40.523 "job": "raid_bdev1", 00:09:40.523 "core_mask": "0x1", 00:09:40.523 "workload": "randrw", 00:09:40.523 "percentage": 50, 00:09:40.523 "status": "finished", 00:09:40.523 "queue_depth": 1, 00:09:40.523 "io_size": 131072, 00:09:40.523 "runtime": 1.397276, 00:09:40.523 "iops": 14213.369441685107, 00:09:40.523 "mibps": 1776.6711802106383, 00:09:40.523 "io_failed": 1, 00:09:40.523 "io_timeout": 0, 00:09:40.523 "avg_latency_us": 98.8820698615201, 00:09:40.523 "min_latency_us": 25.7117903930131, 00:09:40.523 "max_latency_us": 1430.9170305676855 00:09:40.523 } 00:09:40.523 ], 00:09:40.523 "core_count": 1 00:09:40.523 } 00:09:40.523 15:15:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:40.523 15:15:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 76475 00:09:40.523 15:15:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 76475 ']' 00:09:40.523 15:15:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 76475 00:09:40.523 15:15:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:09:40.523 15:15:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:40.523 15:15:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 76475 00:09:40.523 killing process with pid 76475 00:09:40.523 15:15:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:40.523 15:15:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:40.523 15:15:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 76475' 00:09:40.523 15:15:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 76475 00:09:40.524 [2024-11-19 15:15:30.617147] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:40.524 15:15:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 76475 00:09:40.524 [2024-11-19 15:15:30.665270] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:40.784 15:15:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.4JsUMAgajk 00:09:40.784 15:15:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:09:40.784 15:15:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:09:40.784 15:15:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.72 00:09:40.784 15:15:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:09:40.784 15:15:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:40.784 15:15:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:40.784 15:15:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.72 != \0\.\0\0 ]] 00:09:40.784 00:09:40.784 real 0m3.474s 00:09:40.784 user 0m4.356s 00:09:40.784 sys 0m0.589s 00:09:40.784 15:15:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:40.784 15:15:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.784 ************************************ 00:09:40.784 END TEST raid_read_error_test 00:09:40.784 ************************************ 00:09:40.784 15:15:31 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid0 3 write 00:09:40.784 15:15:31 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:40.784 15:15:31 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:40.784 15:15:31 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:40.784 ************************************ 00:09:40.784 START TEST raid_write_error_test 00:09:40.784 ************************************ 00:09:40.784 15:15:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid0 3 write 00:09:40.784 15:15:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:09:40.784 15:15:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:09:40.784 15:15:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:09:40.784 15:15:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:09:40.784 15:15:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:40.784 15:15:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:09:40.784 15:15:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:40.784 15:15:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:40.784 15:15:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:09:40.784 15:15:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:40.784 15:15:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:40.784 15:15:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:09:40.784 15:15:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:40.784 15:15:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:40.784 15:15:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:09:40.784 15:15:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:09:40.784 15:15:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:09:40.784 15:15:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:09:40.784 15:15:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:09:40.784 15:15:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:09:40.784 15:15:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:09:40.784 15:15:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:09:40.784 15:15:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:09:40.784 15:15:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:09:40.784 15:15:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:09:40.784 15:15:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.yA7dpsT78A 00:09:40.784 15:15:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=76609 00:09:40.784 15:15:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:09:40.784 15:15:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 76609 00:09:40.784 15:15:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 76609 ']' 00:09:40.784 15:15:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:40.784 15:15:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:40.784 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:40.784 15:15:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:40.784 15:15:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:40.784 15:15:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.044 [2024-11-19 15:15:31.196263] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:09:41.044 [2024-11-19 15:15:31.196425] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid76609 ] 00:09:41.044 [2024-11-19 15:15:31.360753] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:41.303 [2024-11-19 15:15:31.404355] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:41.303 [2024-11-19 15:15:31.481055] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:41.303 [2024-11-19 15:15:31.481094] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:41.871 15:15:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:41.871 15:15:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:09:41.871 15:15:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:41.871 15:15:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:09:41.871 15:15:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.871 15:15:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.871 BaseBdev1_malloc 00:09:41.871 15:15:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.871 15:15:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:09:41.871 15:15:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.871 15:15:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.871 true 00:09:41.871 15:15:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.871 15:15:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:09:41.871 15:15:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.871 15:15:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.871 [2024-11-19 15:15:32.079969] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:09:41.871 [2024-11-19 15:15:32.080043] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:41.871 [2024-11-19 15:15:32.080066] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:09:41.872 [2024-11-19 15:15:32.080075] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:41.872 [2024-11-19 15:15:32.082571] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:41.872 [2024-11-19 15:15:32.082606] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:09:41.872 BaseBdev1 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.872 BaseBdev2_malloc 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.872 true 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.872 [2024-11-19 15:15:32.126833] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:09:41.872 [2024-11-19 15:15:32.126889] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:41.872 [2024-11-19 15:15:32.126923] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:09:41.872 [2024-11-19 15:15:32.126943] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:41.872 [2024-11-19 15:15:32.129487] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:41.872 [2024-11-19 15:15:32.129525] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:09:41.872 BaseBdev2 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.872 BaseBdev3_malloc 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.872 true 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.872 [2024-11-19 15:15:32.173630] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:09:41.872 [2024-11-19 15:15:32.173683] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:41.872 [2024-11-19 15:15:32.173702] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:09:41.872 [2024-11-19 15:15:32.173711] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:41.872 [2024-11-19 15:15:32.176177] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:41.872 [2024-11-19 15:15:32.176212] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:09:41.872 BaseBdev3 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.872 [2024-11-19 15:15:32.185699] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:41.872 [2024-11-19 15:15:32.187863] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:41.872 [2024-11-19 15:15:32.187937] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:41.872 [2024-11-19 15:15:32.188150] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:09:41.872 [2024-11-19 15:15:32.188171] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:09:41.872 [2024-11-19 15:15:32.188436] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002bb0 00:09:41.872 [2024-11-19 15:15:32.188576] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:09:41.872 [2024-11-19 15:15:32.188590] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:09:41.872 [2024-11-19 15:15:32.188723] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.872 15:15:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.131 15:15:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:42.131 15:15:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:42.131 "name": "raid_bdev1", 00:09:42.131 "uuid": "6e649885-15d9-4815-ba97-b37785055a8b", 00:09:42.131 "strip_size_kb": 64, 00:09:42.131 "state": "online", 00:09:42.131 "raid_level": "raid0", 00:09:42.131 "superblock": true, 00:09:42.131 "num_base_bdevs": 3, 00:09:42.131 "num_base_bdevs_discovered": 3, 00:09:42.131 "num_base_bdevs_operational": 3, 00:09:42.131 "base_bdevs_list": [ 00:09:42.131 { 00:09:42.131 "name": "BaseBdev1", 00:09:42.131 "uuid": "76a2c8c8-b5bc-5291-9601-c9ab9d5242a4", 00:09:42.132 "is_configured": true, 00:09:42.132 "data_offset": 2048, 00:09:42.132 "data_size": 63488 00:09:42.132 }, 00:09:42.132 { 00:09:42.132 "name": "BaseBdev2", 00:09:42.132 "uuid": "d53a0214-ad76-5720-a406-215254b41670", 00:09:42.132 "is_configured": true, 00:09:42.132 "data_offset": 2048, 00:09:42.132 "data_size": 63488 00:09:42.132 }, 00:09:42.132 { 00:09:42.132 "name": "BaseBdev3", 00:09:42.132 "uuid": "2b3603d6-c549-5e14-8e5f-8122e4ffca58", 00:09:42.132 "is_configured": true, 00:09:42.132 "data_offset": 2048, 00:09:42.132 "data_size": 63488 00:09:42.132 } 00:09:42.132 ] 00:09:42.132 }' 00:09:42.132 15:15:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:42.132 15:15:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.391 15:15:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:09:42.391 15:15:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:09:42.391 [2024-11-19 15:15:32.709358] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002d50 00:09:43.332 15:15:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:09:43.332 15:15:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.332 15:15:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.332 15:15:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.332 15:15:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:09:43.332 15:15:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:09:43.332 15:15:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:09:43.332 15:15:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:09:43.332 15:15:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:43.332 15:15:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:43.332 15:15:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:43.332 15:15:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:43.332 15:15:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:43.332 15:15:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:43.332 15:15:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:43.332 15:15:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:43.332 15:15:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:43.332 15:15:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:43.332 15:15:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:43.332 15:15:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.332 15:15:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.332 15:15:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.592 15:15:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:43.592 "name": "raid_bdev1", 00:09:43.592 "uuid": "6e649885-15d9-4815-ba97-b37785055a8b", 00:09:43.592 "strip_size_kb": 64, 00:09:43.592 "state": "online", 00:09:43.592 "raid_level": "raid0", 00:09:43.592 "superblock": true, 00:09:43.592 "num_base_bdevs": 3, 00:09:43.592 "num_base_bdevs_discovered": 3, 00:09:43.592 "num_base_bdevs_operational": 3, 00:09:43.592 "base_bdevs_list": [ 00:09:43.592 { 00:09:43.592 "name": "BaseBdev1", 00:09:43.592 "uuid": "76a2c8c8-b5bc-5291-9601-c9ab9d5242a4", 00:09:43.592 "is_configured": true, 00:09:43.592 "data_offset": 2048, 00:09:43.592 "data_size": 63488 00:09:43.592 }, 00:09:43.592 { 00:09:43.592 "name": "BaseBdev2", 00:09:43.592 "uuid": "d53a0214-ad76-5720-a406-215254b41670", 00:09:43.592 "is_configured": true, 00:09:43.592 "data_offset": 2048, 00:09:43.592 "data_size": 63488 00:09:43.592 }, 00:09:43.592 { 00:09:43.592 "name": "BaseBdev3", 00:09:43.592 "uuid": "2b3603d6-c549-5e14-8e5f-8122e4ffca58", 00:09:43.592 "is_configured": true, 00:09:43.592 "data_offset": 2048, 00:09:43.592 "data_size": 63488 00:09:43.592 } 00:09:43.592 ] 00:09:43.592 }' 00:09:43.592 15:15:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:43.592 15:15:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.853 15:15:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:43.853 15:15:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.853 15:15:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.853 [2024-11-19 15:15:34.074350] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:43.853 [2024-11-19 15:15:34.074398] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:43.853 [2024-11-19 15:15:34.077081] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:43.853 [2024-11-19 15:15:34.077152] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:43.853 [2024-11-19 15:15:34.077196] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:43.853 [2024-11-19 15:15:34.077208] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:09:43.853 { 00:09:43.853 "results": [ 00:09:43.853 { 00:09:43.853 "job": "raid_bdev1", 00:09:43.853 "core_mask": "0x1", 00:09:43.853 "workload": "randrw", 00:09:43.853 "percentage": 50, 00:09:43.853 "status": "finished", 00:09:43.853 "queue_depth": 1, 00:09:43.853 "io_size": 131072, 00:09:43.853 "runtime": 1.365452, 00:09:43.853 "iops": 14324.926837413544, 00:09:43.853 "mibps": 1790.615854676693, 00:09:43.853 "io_failed": 1, 00:09:43.853 "io_timeout": 0, 00:09:43.853 "avg_latency_us": 98.14896756736123, 00:09:43.853 "min_latency_us": 21.351965065502185, 00:09:43.853 "max_latency_us": 1330.7528384279476 00:09:43.853 } 00:09:43.853 ], 00:09:43.853 "core_count": 1 00:09:43.853 } 00:09:43.853 15:15:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.853 15:15:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 76609 00:09:43.853 15:15:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 76609 ']' 00:09:43.853 15:15:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 76609 00:09:43.853 15:15:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:09:43.853 15:15:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:43.853 15:15:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 76609 00:09:43.853 15:15:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:43.853 15:15:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:43.853 15:15:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 76609' 00:09:43.853 killing process with pid 76609 00:09:43.853 15:15:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 76609 00:09:43.853 [2024-11-19 15:15:34.126053] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:43.853 15:15:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 76609 00:09:43.853 [2024-11-19 15:15:34.175081] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:44.420 15:15:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:09:44.420 15:15:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.yA7dpsT78A 00:09:44.420 15:15:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:09:44.420 15:15:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.73 00:09:44.420 15:15:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:09:44.420 15:15:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:44.420 15:15:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:44.420 15:15:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.73 != \0\.\0\0 ]] 00:09:44.420 00:09:44.420 real 0m3.438s 00:09:44.420 user 0m4.253s 00:09:44.420 sys 0m0.615s 00:09:44.420 15:15:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:44.420 15:15:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.420 ************************************ 00:09:44.420 END TEST raid_write_error_test 00:09:44.420 ************************************ 00:09:44.420 15:15:34 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:09:44.420 15:15:34 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test concat 3 false 00:09:44.420 15:15:34 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:44.420 15:15:34 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:44.420 15:15:34 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:44.420 ************************************ 00:09:44.420 START TEST raid_state_function_test 00:09:44.420 ************************************ 00:09:44.420 15:15:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test concat 3 false 00:09:44.420 15:15:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:09:44.420 15:15:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:09:44.420 15:15:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:09:44.420 15:15:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:09:44.420 15:15:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:09:44.420 15:15:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:44.420 15:15:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:09:44.420 15:15:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:44.420 15:15:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:44.420 15:15:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:09:44.420 15:15:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:44.420 15:15:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:44.420 15:15:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:09:44.420 15:15:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:44.420 15:15:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:44.420 15:15:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:09:44.420 15:15:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:09:44.420 15:15:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:09:44.420 15:15:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:09:44.420 15:15:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:09:44.420 15:15:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:09:44.420 15:15:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:09:44.420 15:15:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:09:44.420 15:15:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:09:44.420 15:15:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:09:44.420 15:15:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:09:44.420 15:15:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=76742 00:09:44.420 15:15:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:09:44.420 15:15:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 76742' 00:09:44.420 Process raid pid: 76742 00:09:44.420 15:15:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 76742 00:09:44.420 15:15:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 76742 ']' 00:09:44.420 15:15:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:44.420 15:15:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:44.420 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:44.420 15:15:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:44.420 15:15:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:44.420 15:15:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.420 [2024-11-19 15:15:34.670214] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:09:44.421 [2024-11-19 15:15:34.670334] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:44.682 [2024-11-19 15:15:34.827659] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:44.682 [2024-11-19 15:15:34.868875] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:44.682 [2024-11-19 15:15:34.944750] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:44.682 [2024-11-19 15:15:34.944792] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:45.265 15:15:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:45.265 15:15:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:09:45.265 15:15:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:45.265 15:15:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:45.265 15:15:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.265 [2024-11-19 15:15:35.508046] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:45.265 [2024-11-19 15:15:35.508117] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:45.265 [2024-11-19 15:15:35.508129] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:45.265 [2024-11-19 15:15:35.508140] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:45.265 [2024-11-19 15:15:35.508148] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:45.265 [2024-11-19 15:15:35.508161] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:45.265 15:15:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:45.265 15:15:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:45.265 15:15:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:45.265 15:15:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:45.265 15:15:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:45.265 15:15:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:45.265 15:15:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:45.265 15:15:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:45.265 15:15:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:45.265 15:15:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:45.265 15:15:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:45.265 15:15:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:45.265 15:15:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:45.265 15:15:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:45.265 15:15:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.265 15:15:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:45.265 15:15:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:45.265 "name": "Existed_Raid", 00:09:45.265 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:45.265 "strip_size_kb": 64, 00:09:45.265 "state": "configuring", 00:09:45.265 "raid_level": "concat", 00:09:45.265 "superblock": false, 00:09:45.265 "num_base_bdevs": 3, 00:09:45.265 "num_base_bdevs_discovered": 0, 00:09:45.265 "num_base_bdevs_operational": 3, 00:09:45.265 "base_bdevs_list": [ 00:09:45.265 { 00:09:45.265 "name": "BaseBdev1", 00:09:45.265 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:45.265 "is_configured": false, 00:09:45.265 "data_offset": 0, 00:09:45.265 "data_size": 0 00:09:45.265 }, 00:09:45.265 { 00:09:45.265 "name": "BaseBdev2", 00:09:45.265 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:45.265 "is_configured": false, 00:09:45.265 "data_offset": 0, 00:09:45.265 "data_size": 0 00:09:45.265 }, 00:09:45.265 { 00:09:45.265 "name": "BaseBdev3", 00:09:45.265 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:45.265 "is_configured": false, 00:09:45.265 "data_offset": 0, 00:09:45.265 "data_size": 0 00:09:45.265 } 00:09:45.265 ] 00:09:45.265 }' 00:09:45.265 15:15:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:45.265 15:15:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.836 [2024-11-19 15:15:36.006971] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:45.836 [2024-11-19 15:15:36.007048] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.836 [2024-11-19 15:15:36.014960] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:45.836 [2024-11-19 15:15:36.015026] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:45.836 [2024-11-19 15:15:36.015035] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:45.836 [2024-11-19 15:15:36.015045] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:45.836 [2024-11-19 15:15:36.015052] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:45.836 [2024-11-19 15:15:36.015061] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.836 [2024-11-19 15:15:36.038295] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:45.836 BaseBdev1 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.836 [ 00:09:45.836 { 00:09:45.836 "name": "BaseBdev1", 00:09:45.836 "aliases": [ 00:09:45.836 "1b6a4a6b-9209-4529-979c-392ed5ba2470" 00:09:45.836 ], 00:09:45.836 "product_name": "Malloc disk", 00:09:45.836 "block_size": 512, 00:09:45.836 "num_blocks": 65536, 00:09:45.836 "uuid": "1b6a4a6b-9209-4529-979c-392ed5ba2470", 00:09:45.836 "assigned_rate_limits": { 00:09:45.836 "rw_ios_per_sec": 0, 00:09:45.836 "rw_mbytes_per_sec": 0, 00:09:45.836 "r_mbytes_per_sec": 0, 00:09:45.836 "w_mbytes_per_sec": 0 00:09:45.836 }, 00:09:45.836 "claimed": true, 00:09:45.836 "claim_type": "exclusive_write", 00:09:45.836 "zoned": false, 00:09:45.836 "supported_io_types": { 00:09:45.836 "read": true, 00:09:45.836 "write": true, 00:09:45.836 "unmap": true, 00:09:45.836 "flush": true, 00:09:45.836 "reset": true, 00:09:45.836 "nvme_admin": false, 00:09:45.836 "nvme_io": false, 00:09:45.836 "nvme_io_md": false, 00:09:45.836 "write_zeroes": true, 00:09:45.836 "zcopy": true, 00:09:45.836 "get_zone_info": false, 00:09:45.836 "zone_management": false, 00:09:45.836 "zone_append": false, 00:09:45.836 "compare": false, 00:09:45.836 "compare_and_write": false, 00:09:45.836 "abort": true, 00:09:45.836 "seek_hole": false, 00:09:45.836 "seek_data": false, 00:09:45.836 "copy": true, 00:09:45.836 "nvme_iov_md": false 00:09:45.836 }, 00:09:45.836 "memory_domains": [ 00:09:45.836 { 00:09:45.836 "dma_device_id": "system", 00:09:45.836 "dma_device_type": 1 00:09:45.836 }, 00:09:45.836 { 00:09:45.836 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:45.836 "dma_device_type": 2 00:09:45.836 } 00:09:45.836 ], 00:09:45.836 "driver_specific": {} 00:09:45.836 } 00:09:45.836 ] 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:45.836 "name": "Existed_Raid", 00:09:45.836 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:45.836 "strip_size_kb": 64, 00:09:45.836 "state": "configuring", 00:09:45.836 "raid_level": "concat", 00:09:45.836 "superblock": false, 00:09:45.836 "num_base_bdevs": 3, 00:09:45.836 "num_base_bdevs_discovered": 1, 00:09:45.836 "num_base_bdevs_operational": 3, 00:09:45.836 "base_bdevs_list": [ 00:09:45.836 { 00:09:45.836 "name": "BaseBdev1", 00:09:45.836 "uuid": "1b6a4a6b-9209-4529-979c-392ed5ba2470", 00:09:45.836 "is_configured": true, 00:09:45.836 "data_offset": 0, 00:09:45.836 "data_size": 65536 00:09:45.836 }, 00:09:45.836 { 00:09:45.836 "name": "BaseBdev2", 00:09:45.836 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:45.836 "is_configured": false, 00:09:45.836 "data_offset": 0, 00:09:45.836 "data_size": 0 00:09:45.836 }, 00:09:45.836 { 00:09:45.836 "name": "BaseBdev3", 00:09:45.836 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:45.836 "is_configured": false, 00:09:45.836 "data_offset": 0, 00:09:45.836 "data_size": 0 00:09:45.836 } 00:09:45.836 ] 00:09:45.836 }' 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:45.836 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.406 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:46.406 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:46.406 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.406 [2024-11-19 15:15:36.441697] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:46.406 [2024-11-19 15:15:36.441778] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:09:46.406 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:46.406 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:46.406 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:46.406 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.406 [2024-11-19 15:15:36.453690] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:46.406 [2024-11-19 15:15:36.455896] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:46.406 [2024-11-19 15:15:36.455943] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:46.406 [2024-11-19 15:15:36.455953] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:46.406 [2024-11-19 15:15:36.455992] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:46.406 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:46.406 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:09:46.406 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:46.406 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:46.406 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:46.406 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:46.406 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:46.406 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:46.406 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:46.406 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:46.406 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:46.406 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:46.406 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:46.406 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:46.406 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:46.406 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.406 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:46.406 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:46.406 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:46.406 "name": "Existed_Raid", 00:09:46.406 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:46.406 "strip_size_kb": 64, 00:09:46.406 "state": "configuring", 00:09:46.406 "raid_level": "concat", 00:09:46.406 "superblock": false, 00:09:46.406 "num_base_bdevs": 3, 00:09:46.406 "num_base_bdevs_discovered": 1, 00:09:46.406 "num_base_bdevs_operational": 3, 00:09:46.406 "base_bdevs_list": [ 00:09:46.406 { 00:09:46.406 "name": "BaseBdev1", 00:09:46.406 "uuid": "1b6a4a6b-9209-4529-979c-392ed5ba2470", 00:09:46.406 "is_configured": true, 00:09:46.406 "data_offset": 0, 00:09:46.406 "data_size": 65536 00:09:46.406 }, 00:09:46.406 { 00:09:46.407 "name": "BaseBdev2", 00:09:46.407 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:46.407 "is_configured": false, 00:09:46.407 "data_offset": 0, 00:09:46.407 "data_size": 0 00:09:46.407 }, 00:09:46.407 { 00:09:46.407 "name": "BaseBdev3", 00:09:46.407 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:46.407 "is_configured": false, 00:09:46.407 "data_offset": 0, 00:09:46.407 "data_size": 0 00:09:46.407 } 00:09:46.407 ] 00:09:46.407 }' 00:09:46.407 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:46.407 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.667 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:46.667 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:46.667 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.667 [2024-11-19 15:15:36.885814] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:46.667 BaseBdev2 00:09:46.667 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:46.667 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:09:46.667 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:46.667 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:46.667 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:46.667 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:46.667 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:46.667 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:46.667 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:46.667 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.667 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:46.667 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:46.667 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:46.667 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.667 [ 00:09:46.667 { 00:09:46.667 "name": "BaseBdev2", 00:09:46.667 "aliases": [ 00:09:46.667 "2988ca0b-1155-406e-bc56-17f324c384fc" 00:09:46.667 ], 00:09:46.667 "product_name": "Malloc disk", 00:09:46.667 "block_size": 512, 00:09:46.667 "num_blocks": 65536, 00:09:46.667 "uuid": "2988ca0b-1155-406e-bc56-17f324c384fc", 00:09:46.667 "assigned_rate_limits": { 00:09:46.667 "rw_ios_per_sec": 0, 00:09:46.667 "rw_mbytes_per_sec": 0, 00:09:46.667 "r_mbytes_per_sec": 0, 00:09:46.667 "w_mbytes_per_sec": 0 00:09:46.667 }, 00:09:46.667 "claimed": true, 00:09:46.667 "claim_type": "exclusive_write", 00:09:46.667 "zoned": false, 00:09:46.667 "supported_io_types": { 00:09:46.667 "read": true, 00:09:46.667 "write": true, 00:09:46.667 "unmap": true, 00:09:46.667 "flush": true, 00:09:46.667 "reset": true, 00:09:46.667 "nvme_admin": false, 00:09:46.667 "nvme_io": false, 00:09:46.667 "nvme_io_md": false, 00:09:46.667 "write_zeroes": true, 00:09:46.667 "zcopy": true, 00:09:46.667 "get_zone_info": false, 00:09:46.667 "zone_management": false, 00:09:46.667 "zone_append": false, 00:09:46.667 "compare": false, 00:09:46.667 "compare_and_write": false, 00:09:46.667 "abort": true, 00:09:46.667 "seek_hole": false, 00:09:46.667 "seek_data": false, 00:09:46.667 "copy": true, 00:09:46.667 "nvme_iov_md": false 00:09:46.667 }, 00:09:46.667 "memory_domains": [ 00:09:46.667 { 00:09:46.667 "dma_device_id": "system", 00:09:46.667 "dma_device_type": 1 00:09:46.667 }, 00:09:46.667 { 00:09:46.667 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:46.667 "dma_device_type": 2 00:09:46.667 } 00:09:46.667 ], 00:09:46.667 "driver_specific": {} 00:09:46.667 } 00:09:46.667 ] 00:09:46.667 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:46.667 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:46.667 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:46.667 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:46.667 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:46.667 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:46.667 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:46.667 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:46.667 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:46.667 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:46.667 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:46.667 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:46.667 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:46.667 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:46.667 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:46.668 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:46.668 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:46.668 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.668 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:46.668 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:46.668 "name": "Existed_Raid", 00:09:46.668 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:46.668 "strip_size_kb": 64, 00:09:46.668 "state": "configuring", 00:09:46.668 "raid_level": "concat", 00:09:46.668 "superblock": false, 00:09:46.668 "num_base_bdevs": 3, 00:09:46.668 "num_base_bdevs_discovered": 2, 00:09:46.668 "num_base_bdevs_operational": 3, 00:09:46.668 "base_bdevs_list": [ 00:09:46.668 { 00:09:46.668 "name": "BaseBdev1", 00:09:46.668 "uuid": "1b6a4a6b-9209-4529-979c-392ed5ba2470", 00:09:46.668 "is_configured": true, 00:09:46.668 "data_offset": 0, 00:09:46.668 "data_size": 65536 00:09:46.668 }, 00:09:46.668 { 00:09:46.668 "name": "BaseBdev2", 00:09:46.668 "uuid": "2988ca0b-1155-406e-bc56-17f324c384fc", 00:09:46.668 "is_configured": true, 00:09:46.668 "data_offset": 0, 00:09:46.668 "data_size": 65536 00:09:46.668 }, 00:09:46.668 { 00:09:46.668 "name": "BaseBdev3", 00:09:46.668 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:46.668 "is_configured": false, 00:09:46.668 "data_offset": 0, 00:09:46.668 "data_size": 0 00:09:46.668 } 00:09:46.668 ] 00:09:46.668 }' 00:09:46.668 15:15:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:46.668 15:15:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.237 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:47.237 15:15:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:47.237 15:15:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.237 [2024-11-19 15:15:37.348180] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:47.237 [2024-11-19 15:15:37.348232] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:09:47.237 [2024-11-19 15:15:37.348244] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:09:47.237 [2024-11-19 15:15:37.348559] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:09:47.237 [2024-11-19 15:15:37.348723] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:09:47.237 [2024-11-19 15:15:37.348734] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:09:47.237 [2024-11-19 15:15:37.348986] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:47.237 BaseBdev3 00:09:47.237 15:15:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:47.237 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:09:47.237 15:15:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:09:47.237 15:15:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:47.237 15:15:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:47.237 15:15:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:47.237 15:15:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:47.237 15:15:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:47.237 15:15:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:47.237 15:15:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.237 15:15:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:47.237 15:15:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:47.237 15:15:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:47.237 15:15:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.237 [ 00:09:47.237 { 00:09:47.237 "name": "BaseBdev3", 00:09:47.237 "aliases": [ 00:09:47.237 "41ac5304-1955-4db3-b427-387d8ea5f78f" 00:09:47.237 ], 00:09:47.237 "product_name": "Malloc disk", 00:09:47.237 "block_size": 512, 00:09:47.238 "num_blocks": 65536, 00:09:47.238 "uuid": "41ac5304-1955-4db3-b427-387d8ea5f78f", 00:09:47.238 "assigned_rate_limits": { 00:09:47.238 "rw_ios_per_sec": 0, 00:09:47.238 "rw_mbytes_per_sec": 0, 00:09:47.238 "r_mbytes_per_sec": 0, 00:09:47.238 "w_mbytes_per_sec": 0 00:09:47.238 }, 00:09:47.238 "claimed": true, 00:09:47.238 "claim_type": "exclusive_write", 00:09:47.238 "zoned": false, 00:09:47.238 "supported_io_types": { 00:09:47.238 "read": true, 00:09:47.238 "write": true, 00:09:47.238 "unmap": true, 00:09:47.238 "flush": true, 00:09:47.238 "reset": true, 00:09:47.238 "nvme_admin": false, 00:09:47.238 "nvme_io": false, 00:09:47.238 "nvme_io_md": false, 00:09:47.238 "write_zeroes": true, 00:09:47.238 "zcopy": true, 00:09:47.238 "get_zone_info": false, 00:09:47.238 "zone_management": false, 00:09:47.238 "zone_append": false, 00:09:47.238 "compare": false, 00:09:47.238 "compare_and_write": false, 00:09:47.238 "abort": true, 00:09:47.238 "seek_hole": false, 00:09:47.238 "seek_data": false, 00:09:47.238 "copy": true, 00:09:47.238 "nvme_iov_md": false 00:09:47.238 }, 00:09:47.238 "memory_domains": [ 00:09:47.238 { 00:09:47.238 "dma_device_id": "system", 00:09:47.238 "dma_device_type": 1 00:09:47.238 }, 00:09:47.238 { 00:09:47.238 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:47.238 "dma_device_type": 2 00:09:47.238 } 00:09:47.238 ], 00:09:47.238 "driver_specific": {} 00:09:47.238 } 00:09:47.238 ] 00:09:47.238 15:15:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:47.238 15:15:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:47.238 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:47.238 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:47.238 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:09:47.238 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:47.238 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:47.238 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:47.238 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:47.238 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:47.238 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:47.238 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:47.238 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:47.238 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:47.238 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:47.238 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:47.238 15:15:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:47.238 15:15:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.238 15:15:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:47.238 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:47.238 "name": "Existed_Raid", 00:09:47.238 "uuid": "7a21c696-b71e-48ef-a425-16d64cd39465", 00:09:47.238 "strip_size_kb": 64, 00:09:47.238 "state": "online", 00:09:47.238 "raid_level": "concat", 00:09:47.238 "superblock": false, 00:09:47.238 "num_base_bdevs": 3, 00:09:47.238 "num_base_bdevs_discovered": 3, 00:09:47.238 "num_base_bdevs_operational": 3, 00:09:47.238 "base_bdevs_list": [ 00:09:47.238 { 00:09:47.238 "name": "BaseBdev1", 00:09:47.238 "uuid": "1b6a4a6b-9209-4529-979c-392ed5ba2470", 00:09:47.238 "is_configured": true, 00:09:47.238 "data_offset": 0, 00:09:47.238 "data_size": 65536 00:09:47.238 }, 00:09:47.238 { 00:09:47.238 "name": "BaseBdev2", 00:09:47.238 "uuid": "2988ca0b-1155-406e-bc56-17f324c384fc", 00:09:47.238 "is_configured": true, 00:09:47.238 "data_offset": 0, 00:09:47.238 "data_size": 65536 00:09:47.238 }, 00:09:47.238 { 00:09:47.238 "name": "BaseBdev3", 00:09:47.238 "uuid": "41ac5304-1955-4db3-b427-387d8ea5f78f", 00:09:47.238 "is_configured": true, 00:09:47.238 "data_offset": 0, 00:09:47.238 "data_size": 65536 00:09:47.238 } 00:09:47.238 ] 00:09:47.238 }' 00:09:47.238 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:47.238 15:15:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.498 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:09:47.498 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:47.498 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:47.498 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:47.498 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:47.498 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:47.498 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:47.498 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:47.498 15:15:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:47.498 15:15:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.498 [2024-11-19 15:15:37.779865] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:47.498 15:15:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:47.498 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:47.498 "name": "Existed_Raid", 00:09:47.498 "aliases": [ 00:09:47.498 "7a21c696-b71e-48ef-a425-16d64cd39465" 00:09:47.498 ], 00:09:47.498 "product_name": "Raid Volume", 00:09:47.498 "block_size": 512, 00:09:47.498 "num_blocks": 196608, 00:09:47.498 "uuid": "7a21c696-b71e-48ef-a425-16d64cd39465", 00:09:47.498 "assigned_rate_limits": { 00:09:47.498 "rw_ios_per_sec": 0, 00:09:47.498 "rw_mbytes_per_sec": 0, 00:09:47.498 "r_mbytes_per_sec": 0, 00:09:47.498 "w_mbytes_per_sec": 0 00:09:47.498 }, 00:09:47.498 "claimed": false, 00:09:47.498 "zoned": false, 00:09:47.498 "supported_io_types": { 00:09:47.498 "read": true, 00:09:47.498 "write": true, 00:09:47.498 "unmap": true, 00:09:47.498 "flush": true, 00:09:47.498 "reset": true, 00:09:47.498 "nvme_admin": false, 00:09:47.498 "nvme_io": false, 00:09:47.498 "nvme_io_md": false, 00:09:47.498 "write_zeroes": true, 00:09:47.498 "zcopy": false, 00:09:47.498 "get_zone_info": false, 00:09:47.498 "zone_management": false, 00:09:47.498 "zone_append": false, 00:09:47.498 "compare": false, 00:09:47.498 "compare_and_write": false, 00:09:47.498 "abort": false, 00:09:47.498 "seek_hole": false, 00:09:47.498 "seek_data": false, 00:09:47.498 "copy": false, 00:09:47.498 "nvme_iov_md": false 00:09:47.498 }, 00:09:47.498 "memory_domains": [ 00:09:47.498 { 00:09:47.498 "dma_device_id": "system", 00:09:47.498 "dma_device_type": 1 00:09:47.498 }, 00:09:47.498 { 00:09:47.498 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:47.498 "dma_device_type": 2 00:09:47.498 }, 00:09:47.498 { 00:09:47.498 "dma_device_id": "system", 00:09:47.498 "dma_device_type": 1 00:09:47.498 }, 00:09:47.498 { 00:09:47.498 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:47.498 "dma_device_type": 2 00:09:47.498 }, 00:09:47.498 { 00:09:47.498 "dma_device_id": "system", 00:09:47.498 "dma_device_type": 1 00:09:47.498 }, 00:09:47.498 { 00:09:47.498 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:47.498 "dma_device_type": 2 00:09:47.498 } 00:09:47.498 ], 00:09:47.498 "driver_specific": { 00:09:47.498 "raid": { 00:09:47.498 "uuid": "7a21c696-b71e-48ef-a425-16d64cd39465", 00:09:47.498 "strip_size_kb": 64, 00:09:47.498 "state": "online", 00:09:47.498 "raid_level": "concat", 00:09:47.498 "superblock": false, 00:09:47.498 "num_base_bdevs": 3, 00:09:47.498 "num_base_bdevs_discovered": 3, 00:09:47.498 "num_base_bdevs_operational": 3, 00:09:47.498 "base_bdevs_list": [ 00:09:47.498 { 00:09:47.498 "name": "BaseBdev1", 00:09:47.498 "uuid": "1b6a4a6b-9209-4529-979c-392ed5ba2470", 00:09:47.498 "is_configured": true, 00:09:47.498 "data_offset": 0, 00:09:47.498 "data_size": 65536 00:09:47.498 }, 00:09:47.498 { 00:09:47.498 "name": "BaseBdev2", 00:09:47.498 "uuid": "2988ca0b-1155-406e-bc56-17f324c384fc", 00:09:47.498 "is_configured": true, 00:09:47.498 "data_offset": 0, 00:09:47.498 "data_size": 65536 00:09:47.498 }, 00:09:47.498 { 00:09:47.498 "name": "BaseBdev3", 00:09:47.498 "uuid": "41ac5304-1955-4db3-b427-387d8ea5f78f", 00:09:47.499 "is_configured": true, 00:09:47.499 "data_offset": 0, 00:09:47.499 "data_size": 65536 00:09:47.499 } 00:09:47.499 ] 00:09:47.499 } 00:09:47.499 } 00:09:47.499 }' 00:09:47.499 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:47.758 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:09:47.758 BaseBdev2 00:09:47.758 BaseBdev3' 00:09:47.758 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:47.758 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:47.758 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:47.758 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:09:47.758 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:47.758 15:15:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:47.758 15:15:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.759 15:15:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:47.759 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:47.759 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:47.759 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:47.759 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:47.759 15:15:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:47.759 15:15:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.759 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:47.759 15:15:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:47.759 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:47.759 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:47.759 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:47.759 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:47.759 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:47.759 15:15:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:47.759 15:15:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.759 15:15:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:47.759 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:47.759 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:47.759 15:15:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:47.759 15:15:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:47.759 15:15:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.759 [2024-11-19 15:15:37.999150] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:47.759 [2024-11-19 15:15:37.999232] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:47.759 [2024-11-19 15:15:37.999341] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:47.759 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:47.759 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:09:47.759 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:09:47.759 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:47.759 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:47.759 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:09:47.759 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 2 00:09:47.759 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:47.759 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:09:47.759 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:47.759 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:47.759 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:47.759 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:47.759 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:47.759 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:47.759 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:47.759 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:47.759 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:47.759 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.759 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:47.759 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:47.759 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:47.759 "name": "Existed_Raid", 00:09:47.759 "uuid": "7a21c696-b71e-48ef-a425-16d64cd39465", 00:09:47.759 "strip_size_kb": 64, 00:09:47.759 "state": "offline", 00:09:47.759 "raid_level": "concat", 00:09:47.759 "superblock": false, 00:09:47.759 "num_base_bdevs": 3, 00:09:47.759 "num_base_bdevs_discovered": 2, 00:09:47.759 "num_base_bdevs_operational": 2, 00:09:47.759 "base_bdevs_list": [ 00:09:47.759 { 00:09:47.759 "name": null, 00:09:47.759 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:47.759 "is_configured": false, 00:09:47.759 "data_offset": 0, 00:09:47.759 "data_size": 65536 00:09:47.759 }, 00:09:47.759 { 00:09:47.759 "name": "BaseBdev2", 00:09:47.759 "uuid": "2988ca0b-1155-406e-bc56-17f324c384fc", 00:09:47.759 "is_configured": true, 00:09:47.759 "data_offset": 0, 00:09:47.759 "data_size": 65536 00:09:47.759 }, 00:09:47.759 { 00:09:47.759 "name": "BaseBdev3", 00:09:47.759 "uuid": "41ac5304-1955-4db3-b427-387d8ea5f78f", 00:09:47.759 "is_configured": true, 00:09:47.759 "data_offset": 0, 00:09:47.759 "data_size": 65536 00:09:47.759 } 00:09:47.759 ] 00:09:47.759 }' 00:09:47.759 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:47.759 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.328 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:09:48.328 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:48.328 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:48.328 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:48.328 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.328 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.328 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.328 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:48.328 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:48.328 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:09:48.328 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.328 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.328 [2024-11-19 15:15:38.503158] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:48.328 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.328 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:48.328 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:48.328 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:48.328 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:48.328 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.328 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.328 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.328 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:48.328 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:48.328 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:09:48.328 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.328 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.328 [2024-11-19 15:15:38.572019] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:48.328 [2024-11-19 15:15:38.572091] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:09:48.328 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.328 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:48.328 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:48.328 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:48.328 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:09:48.328 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.328 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.328 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.328 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:09:48.328 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:09:48.328 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:09:48.328 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:09:48.328 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:48.328 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:48.328 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.328 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.588 BaseBdev2 00:09:48.588 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.588 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:09:48.588 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:48.588 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:48.588 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:48.588 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:48.588 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:48.588 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:48.588 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.588 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.588 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.588 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:48.588 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.588 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.588 [ 00:09:48.588 { 00:09:48.588 "name": "BaseBdev2", 00:09:48.588 "aliases": [ 00:09:48.588 "ede0b109-be63-4ba3-bd3f-40824b74fd9e" 00:09:48.588 ], 00:09:48.588 "product_name": "Malloc disk", 00:09:48.588 "block_size": 512, 00:09:48.588 "num_blocks": 65536, 00:09:48.588 "uuid": "ede0b109-be63-4ba3-bd3f-40824b74fd9e", 00:09:48.588 "assigned_rate_limits": { 00:09:48.588 "rw_ios_per_sec": 0, 00:09:48.588 "rw_mbytes_per_sec": 0, 00:09:48.588 "r_mbytes_per_sec": 0, 00:09:48.588 "w_mbytes_per_sec": 0 00:09:48.588 }, 00:09:48.588 "claimed": false, 00:09:48.588 "zoned": false, 00:09:48.588 "supported_io_types": { 00:09:48.588 "read": true, 00:09:48.588 "write": true, 00:09:48.588 "unmap": true, 00:09:48.588 "flush": true, 00:09:48.588 "reset": true, 00:09:48.588 "nvme_admin": false, 00:09:48.588 "nvme_io": false, 00:09:48.588 "nvme_io_md": false, 00:09:48.588 "write_zeroes": true, 00:09:48.588 "zcopy": true, 00:09:48.588 "get_zone_info": false, 00:09:48.588 "zone_management": false, 00:09:48.588 "zone_append": false, 00:09:48.588 "compare": false, 00:09:48.588 "compare_and_write": false, 00:09:48.588 "abort": true, 00:09:48.588 "seek_hole": false, 00:09:48.588 "seek_data": false, 00:09:48.588 "copy": true, 00:09:48.588 "nvme_iov_md": false 00:09:48.588 }, 00:09:48.588 "memory_domains": [ 00:09:48.588 { 00:09:48.588 "dma_device_id": "system", 00:09:48.588 "dma_device_type": 1 00:09:48.588 }, 00:09:48.588 { 00:09:48.588 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:48.588 "dma_device_type": 2 00:09:48.588 } 00:09:48.588 ], 00:09:48.588 "driver_specific": {} 00:09:48.588 } 00:09:48.588 ] 00:09:48.588 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.588 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:48.588 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:48.588 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:48.588 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:48.588 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.588 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.588 BaseBdev3 00:09:48.588 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.588 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:09:48.588 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:09:48.588 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:48.588 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:48.588 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:48.588 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:48.589 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:48.589 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.589 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.589 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.589 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:48.589 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.589 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.589 [ 00:09:48.589 { 00:09:48.589 "name": "BaseBdev3", 00:09:48.589 "aliases": [ 00:09:48.589 "aa2de657-01e9-473e-998c-19cae8d88e03" 00:09:48.589 ], 00:09:48.589 "product_name": "Malloc disk", 00:09:48.589 "block_size": 512, 00:09:48.589 "num_blocks": 65536, 00:09:48.589 "uuid": "aa2de657-01e9-473e-998c-19cae8d88e03", 00:09:48.589 "assigned_rate_limits": { 00:09:48.589 "rw_ios_per_sec": 0, 00:09:48.589 "rw_mbytes_per_sec": 0, 00:09:48.589 "r_mbytes_per_sec": 0, 00:09:48.589 "w_mbytes_per_sec": 0 00:09:48.589 }, 00:09:48.589 "claimed": false, 00:09:48.589 "zoned": false, 00:09:48.589 "supported_io_types": { 00:09:48.589 "read": true, 00:09:48.589 "write": true, 00:09:48.589 "unmap": true, 00:09:48.589 "flush": true, 00:09:48.589 "reset": true, 00:09:48.589 "nvme_admin": false, 00:09:48.589 "nvme_io": false, 00:09:48.589 "nvme_io_md": false, 00:09:48.589 "write_zeroes": true, 00:09:48.589 "zcopy": true, 00:09:48.589 "get_zone_info": false, 00:09:48.589 "zone_management": false, 00:09:48.589 "zone_append": false, 00:09:48.589 "compare": false, 00:09:48.589 "compare_and_write": false, 00:09:48.589 "abort": true, 00:09:48.589 "seek_hole": false, 00:09:48.589 "seek_data": false, 00:09:48.589 "copy": true, 00:09:48.589 "nvme_iov_md": false 00:09:48.589 }, 00:09:48.589 "memory_domains": [ 00:09:48.589 { 00:09:48.589 "dma_device_id": "system", 00:09:48.589 "dma_device_type": 1 00:09:48.589 }, 00:09:48.589 { 00:09:48.589 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:48.589 "dma_device_type": 2 00:09:48.589 } 00:09:48.589 ], 00:09:48.589 "driver_specific": {} 00:09:48.589 } 00:09:48.589 ] 00:09:48.589 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.589 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:48.589 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:48.589 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:48.589 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:48.589 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.589 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.589 [2024-11-19 15:15:38.750823] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:48.589 [2024-11-19 15:15:38.750977] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:48.589 [2024-11-19 15:15:38.751029] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:48.589 [2024-11-19 15:15:38.753312] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:48.589 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.589 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:48.589 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:48.589 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:48.589 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:48.589 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:48.589 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:48.589 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:48.589 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:48.589 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:48.589 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:48.589 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:48.589 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:48.589 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.589 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.589 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.589 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:48.589 "name": "Existed_Raid", 00:09:48.589 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:48.589 "strip_size_kb": 64, 00:09:48.589 "state": "configuring", 00:09:48.589 "raid_level": "concat", 00:09:48.589 "superblock": false, 00:09:48.589 "num_base_bdevs": 3, 00:09:48.589 "num_base_bdevs_discovered": 2, 00:09:48.589 "num_base_bdevs_operational": 3, 00:09:48.589 "base_bdevs_list": [ 00:09:48.589 { 00:09:48.589 "name": "BaseBdev1", 00:09:48.589 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:48.589 "is_configured": false, 00:09:48.589 "data_offset": 0, 00:09:48.589 "data_size": 0 00:09:48.589 }, 00:09:48.589 { 00:09:48.589 "name": "BaseBdev2", 00:09:48.589 "uuid": "ede0b109-be63-4ba3-bd3f-40824b74fd9e", 00:09:48.589 "is_configured": true, 00:09:48.589 "data_offset": 0, 00:09:48.589 "data_size": 65536 00:09:48.589 }, 00:09:48.589 { 00:09:48.589 "name": "BaseBdev3", 00:09:48.589 "uuid": "aa2de657-01e9-473e-998c-19cae8d88e03", 00:09:48.589 "is_configured": true, 00:09:48.589 "data_offset": 0, 00:09:48.589 "data_size": 65536 00:09:48.589 } 00:09:48.589 ] 00:09:48.589 }' 00:09:48.589 15:15:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:48.589 15:15:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.848 15:15:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:09:48.848 15:15:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.107 15:15:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.107 [2024-11-19 15:15:39.190136] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:49.107 15:15:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.107 15:15:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:49.107 15:15:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:49.107 15:15:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:49.107 15:15:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:49.107 15:15:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:49.107 15:15:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:49.107 15:15:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:49.107 15:15:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:49.107 15:15:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:49.107 15:15:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:49.107 15:15:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:49.107 15:15:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:49.107 15:15:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.107 15:15:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.107 15:15:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.107 15:15:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:49.107 "name": "Existed_Raid", 00:09:49.107 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:49.107 "strip_size_kb": 64, 00:09:49.107 "state": "configuring", 00:09:49.107 "raid_level": "concat", 00:09:49.107 "superblock": false, 00:09:49.107 "num_base_bdevs": 3, 00:09:49.107 "num_base_bdevs_discovered": 1, 00:09:49.107 "num_base_bdevs_operational": 3, 00:09:49.107 "base_bdevs_list": [ 00:09:49.107 { 00:09:49.107 "name": "BaseBdev1", 00:09:49.107 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:49.107 "is_configured": false, 00:09:49.107 "data_offset": 0, 00:09:49.107 "data_size": 0 00:09:49.107 }, 00:09:49.107 { 00:09:49.107 "name": null, 00:09:49.107 "uuid": "ede0b109-be63-4ba3-bd3f-40824b74fd9e", 00:09:49.107 "is_configured": false, 00:09:49.107 "data_offset": 0, 00:09:49.107 "data_size": 65536 00:09:49.107 }, 00:09:49.107 { 00:09:49.107 "name": "BaseBdev3", 00:09:49.107 "uuid": "aa2de657-01e9-473e-998c-19cae8d88e03", 00:09:49.107 "is_configured": true, 00:09:49.107 "data_offset": 0, 00:09:49.107 "data_size": 65536 00:09:49.107 } 00:09:49.107 ] 00:09:49.107 }' 00:09:49.107 15:15:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:49.107 15:15:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.366 15:15:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:49.366 15:15:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:49.366 15:15:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.366 15:15:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.366 15:15:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.627 15:15:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:09:49.627 15:15:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:49.627 15:15:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.627 15:15:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.627 [2024-11-19 15:15:39.730069] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:49.627 BaseBdev1 00:09:49.627 15:15:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.627 15:15:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:09:49.627 15:15:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:49.627 15:15:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:49.627 15:15:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:49.627 15:15:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:49.627 15:15:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:49.627 15:15:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:49.627 15:15:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.627 15:15:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.627 15:15:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.627 15:15:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:49.627 15:15:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.627 15:15:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.627 [ 00:09:49.627 { 00:09:49.627 "name": "BaseBdev1", 00:09:49.627 "aliases": [ 00:09:49.627 "5e24f951-48ce-4a67-90a0-0fe4c3cf03d1" 00:09:49.627 ], 00:09:49.627 "product_name": "Malloc disk", 00:09:49.627 "block_size": 512, 00:09:49.627 "num_blocks": 65536, 00:09:49.627 "uuid": "5e24f951-48ce-4a67-90a0-0fe4c3cf03d1", 00:09:49.627 "assigned_rate_limits": { 00:09:49.627 "rw_ios_per_sec": 0, 00:09:49.627 "rw_mbytes_per_sec": 0, 00:09:49.627 "r_mbytes_per_sec": 0, 00:09:49.627 "w_mbytes_per_sec": 0 00:09:49.627 }, 00:09:49.627 "claimed": true, 00:09:49.627 "claim_type": "exclusive_write", 00:09:49.627 "zoned": false, 00:09:49.627 "supported_io_types": { 00:09:49.627 "read": true, 00:09:49.627 "write": true, 00:09:49.627 "unmap": true, 00:09:49.627 "flush": true, 00:09:49.627 "reset": true, 00:09:49.627 "nvme_admin": false, 00:09:49.627 "nvme_io": false, 00:09:49.627 "nvme_io_md": false, 00:09:49.627 "write_zeroes": true, 00:09:49.627 "zcopy": true, 00:09:49.627 "get_zone_info": false, 00:09:49.627 "zone_management": false, 00:09:49.627 "zone_append": false, 00:09:49.627 "compare": false, 00:09:49.627 "compare_and_write": false, 00:09:49.627 "abort": true, 00:09:49.627 "seek_hole": false, 00:09:49.627 "seek_data": false, 00:09:49.627 "copy": true, 00:09:49.627 "nvme_iov_md": false 00:09:49.627 }, 00:09:49.627 "memory_domains": [ 00:09:49.627 { 00:09:49.627 "dma_device_id": "system", 00:09:49.627 "dma_device_type": 1 00:09:49.627 }, 00:09:49.627 { 00:09:49.627 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:49.627 "dma_device_type": 2 00:09:49.627 } 00:09:49.627 ], 00:09:49.627 "driver_specific": {} 00:09:49.627 } 00:09:49.627 ] 00:09:49.627 15:15:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.627 15:15:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:49.627 15:15:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:49.627 15:15:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:49.627 15:15:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:49.627 15:15:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:49.627 15:15:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:49.627 15:15:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:49.627 15:15:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:49.627 15:15:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:49.627 15:15:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:49.627 15:15:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:49.627 15:15:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:49.627 15:15:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:49.627 15:15:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.627 15:15:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.627 15:15:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.627 15:15:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:49.627 "name": "Existed_Raid", 00:09:49.627 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:49.627 "strip_size_kb": 64, 00:09:49.627 "state": "configuring", 00:09:49.627 "raid_level": "concat", 00:09:49.627 "superblock": false, 00:09:49.627 "num_base_bdevs": 3, 00:09:49.627 "num_base_bdevs_discovered": 2, 00:09:49.627 "num_base_bdevs_operational": 3, 00:09:49.627 "base_bdevs_list": [ 00:09:49.627 { 00:09:49.627 "name": "BaseBdev1", 00:09:49.627 "uuid": "5e24f951-48ce-4a67-90a0-0fe4c3cf03d1", 00:09:49.627 "is_configured": true, 00:09:49.627 "data_offset": 0, 00:09:49.627 "data_size": 65536 00:09:49.627 }, 00:09:49.627 { 00:09:49.627 "name": null, 00:09:49.627 "uuid": "ede0b109-be63-4ba3-bd3f-40824b74fd9e", 00:09:49.627 "is_configured": false, 00:09:49.627 "data_offset": 0, 00:09:49.627 "data_size": 65536 00:09:49.627 }, 00:09:49.627 { 00:09:49.627 "name": "BaseBdev3", 00:09:49.627 "uuid": "aa2de657-01e9-473e-998c-19cae8d88e03", 00:09:49.627 "is_configured": true, 00:09:49.627 "data_offset": 0, 00:09:49.627 "data_size": 65536 00:09:49.627 } 00:09:49.627 ] 00:09:49.627 }' 00:09:49.627 15:15:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:49.627 15:15:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.887 15:15:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:49.887 15:15:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:49.887 15:15:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.887 15:15:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.887 15:15:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.887 15:15:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:09:49.887 15:15:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:09:50.146 15:15:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.146 15:15:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:50.146 [2024-11-19 15:15:40.229275] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:50.146 15:15:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.146 15:15:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:50.146 15:15:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:50.146 15:15:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:50.146 15:15:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:50.146 15:15:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:50.146 15:15:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:50.146 15:15:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:50.146 15:15:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:50.146 15:15:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:50.146 15:15:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:50.146 15:15:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:50.146 15:15:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:50.146 15:15:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.146 15:15:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:50.146 15:15:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.146 15:15:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:50.146 "name": "Existed_Raid", 00:09:50.146 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:50.146 "strip_size_kb": 64, 00:09:50.146 "state": "configuring", 00:09:50.146 "raid_level": "concat", 00:09:50.146 "superblock": false, 00:09:50.146 "num_base_bdevs": 3, 00:09:50.146 "num_base_bdevs_discovered": 1, 00:09:50.146 "num_base_bdevs_operational": 3, 00:09:50.146 "base_bdevs_list": [ 00:09:50.146 { 00:09:50.146 "name": "BaseBdev1", 00:09:50.146 "uuid": "5e24f951-48ce-4a67-90a0-0fe4c3cf03d1", 00:09:50.146 "is_configured": true, 00:09:50.146 "data_offset": 0, 00:09:50.146 "data_size": 65536 00:09:50.146 }, 00:09:50.146 { 00:09:50.146 "name": null, 00:09:50.146 "uuid": "ede0b109-be63-4ba3-bd3f-40824b74fd9e", 00:09:50.146 "is_configured": false, 00:09:50.146 "data_offset": 0, 00:09:50.146 "data_size": 65536 00:09:50.146 }, 00:09:50.146 { 00:09:50.146 "name": null, 00:09:50.146 "uuid": "aa2de657-01e9-473e-998c-19cae8d88e03", 00:09:50.146 "is_configured": false, 00:09:50.146 "data_offset": 0, 00:09:50.146 "data_size": 65536 00:09:50.146 } 00:09:50.146 ] 00:09:50.146 }' 00:09:50.146 15:15:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:50.146 15:15:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:50.406 15:15:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:50.406 15:15:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:50.406 15:15:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.406 15:15:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:50.406 15:15:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.406 15:15:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:09:50.406 15:15:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:09:50.406 15:15:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.406 15:15:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:50.406 [2024-11-19 15:15:40.728409] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:50.406 15:15:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.406 15:15:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:50.406 15:15:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:50.406 15:15:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:50.406 15:15:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:50.406 15:15:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:50.406 15:15:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:50.406 15:15:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:50.406 15:15:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:50.406 15:15:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:50.406 15:15:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:50.406 15:15:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:50.406 15:15:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:50.406 15:15:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.406 15:15:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:50.671 15:15:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.671 15:15:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:50.671 "name": "Existed_Raid", 00:09:50.671 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:50.671 "strip_size_kb": 64, 00:09:50.671 "state": "configuring", 00:09:50.671 "raid_level": "concat", 00:09:50.671 "superblock": false, 00:09:50.671 "num_base_bdevs": 3, 00:09:50.671 "num_base_bdevs_discovered": 2, 00:09:50.671 "num_base_bdevs_operational": 3, 00:09:50.671 "base_bdevs_list": [ 00:09:50.671 { 00:09:50.671 "name": "BaseBdev1", 00:09:50.671 "uuid": "5e24f951-48ce-4a67-90a0-0fe4c3cf03d1", 00:09:50.671 "is_configured": true, 00:09:50.671 "data_offset": 0, 00:09:50.671 "data_size": 65536 00:09:50.671 }, 00:09:50.671 { 00:09:50.671 "name": null, 00:09:50.671 "uuid": "ede0b109-be63-4ba3-bd3f-40824b74fd9e", 00:09:50.671 "is_configured": false, 00:09:50.671 "data_offset": 0, 00:09:50.671 "data_size": 65536 00:09:50.671 }, 00:09:50.671 { 00:09:50.671 "name": "BaseBdev3", 00:09:50.671 "uuid": "aa2de657-01e9-473e-998c-19cae8d88e03", 00:09:50.671 "is_configured": true, 00:09:50.671 "data_offset": 0, 00:09:50.671 "data_size": 65536 00:09:50.671 } 00:09:50.671 ] 00:09:50.671 }' 00:09:50.671 15:15:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:50.671 15:15:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:50.940 15:15:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:50.940 15:15:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:50.940 15:15:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.940 15:15:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:50.940 15:15:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.940 15:15:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:09:50.940 15:15:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:50.940 15:15:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.940 15:15:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:50.940 [2024-11-19 15:15:41.215611] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:50.940 15:15:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.940 15:15:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:50.940 15:15:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:50.940 15:15:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:50.940 15:15:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:50.940 15:15:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:50.940 15:15:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:50.940 15:15:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:50.940 15:15:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:50.940 15:15:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:50.940 15:15:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:50.940 15:15:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:50.940 15:15:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:50.940 15:15:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.940 15:15:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:50.940 15:15:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:51.200 15:15:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:51.200 "name": "Existed_Raid", 00:09:51.200 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:51.200 "strip_size_kb": 64, 00:09:51.200 "state": "configuring", 00:09:51.200 "raid_level": "concat", 00:09:51.200 "superblock": false, 00:09:51.200 "num_base_bdevs": 3, 00:09:51.200 "num_base_bdevs_discovered": 1, 00:09:51.200 "num_base_bdevs_operational": 3, 00:09:51.200 "base_bdevs_list": [ 00:09:51.200 { 00:09:51.200 "name": null, 00:09:51.200 "uuid": "5e24f951-48ce-4a67-90a0-0fe4c3cf03d1", 00:09:51.200 "is_configured": false, 00:09:51.200 "data_offset": 0, 00:09:51.200 "data_size": 65536 00:09:51.200 }, 00:09:51.200 { 00:09:51.200 "name": null, 00:09:51.200 "uuid": "ede0b109-be63-4ba3-bd3f-40824b74fd9e", 00:09:51.200 "is_configured": false, 00:09:51.200 "data_offset": 0, 00:09:51.200 "data_size": 65536 00:09:51.200 }, 00:09:51.200 { 00:09:51.200 "name": "BaseBdev3", 00:09:51.200 "uuid": "aa2de657-01e9-473e-998c-19cae8d88e03", 00:09:51.200 "is_configured": true, 00:09:51.200 "data_offset": 0, 00:09:51.200 "data_size": 65536 00:09:51.200 } 00:09:51.200 ] 00:09:51.200 }' 00:09:51.200 15:15:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:51.200 15:15:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.460 15:15:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:51.460 15:15:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:51.460 15:15:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:51.460 15:15:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.460 15:15:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:51.460 15:15:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:09:51.460 15:15:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:09:51.460 15:15:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:51.460 15:15:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.460 [2024-11-19 15:15:41.726932] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:51.460 15:15:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:51.460 15:15:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:51.460 15:15:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:51.460 15:15:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:51.460 15:15:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:51.460 15:15:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:51.460 15:15:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:51.460 15:15:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:51.460 15:15:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:51.460 15:15:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:51.460 15:15:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:51.460 15:15:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:51.460 15:15:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:51.460 15:15:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.460 15:15:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:51.460 15:15:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:51.460 15:15:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:51.460 "name": "Existed_Raid", 00:09:51.460 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:51.460 "strip_size_kb": 64, 00:09:51.460 "state": "configuring", 00:09:51.460 "raid_level": "concat", 00:09:51.460 "superblock": false, 00:09:51.460 "num_base_bdevs": 3, 00:09:51.460 "num_base_bdevs_discovered": 2, 00:09:51.460 "num_base_bdevs_operational": 3, 00:09:51.460 "base_bdevs_list": [ 00:09:51.460 { 00:09:51.460 "name": null, 00:09:51.460 "uuid": "5e24f951-48ce-4a67-90a0-0fe4c3cf03d1", 00:09:51.460 "is_configured": false, 00:09:51.460 "data_offset": 0, 00:09:51.460 "data_size": 65536 00:09:51.460 }, 00:09:51.460 { 00:09:51.460 "name": "BaseBdev2", 00:09:51.460 "uuid": "ede0b109-be63-4ba3-bd3f-40824b74fd9e", 00:09:51.460 "is_configured": true, 00:09:51.460 "data_offset": 0, 00:09:51.460 "data_size": 65536 00:09:51.460 }, 00:09:51.460 { 00:09:51.460 "name": "BaseBdev3", 00:09:51.460 "uuid": "aa2de657-01e9-473e-998c-19cae8d88e03", 00:09:51.460 "is_configured": true, 00:09:51.460 "data_offset": 0, 00:09:51.460 "data_size": 65536 00:09:51.460 } 00:09:51.460 ] 00:09:51.460 }' 00:09:51.460 15:15:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:51.460 15:15:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.028 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:52.028 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:52.028 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.028 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.028 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.028 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:09:52.028 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:52.028 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:09:52.028 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.028 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.028 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.028 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 5e24f951-48ce-4a67-90a0-0fe4c3cf03d1 00:09:52.028 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.028 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.028 [2024-11-19 15:15:42.218939] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:09:52.029 [2024-11-19 15:15:42.219106] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:09:52.029 [2024-11-19 15:15:42.219135] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:09:52.029 [2024-11-19 15:15:42.219449] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:09:52.029 [2024-11-19 15:15:42.219635] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:09:52.029 [2024-11-19 15:15:42.219676] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:09:52.029 [2024-11-19 15:15:42.219948] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:52.029 NewBaseBdev 00:09:52.029 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.029 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:09:52.029 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:09:52.029 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:52.029 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:52.029 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:52.029 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:52.029 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:52.029 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.029 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.029 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.029 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:09:52.029 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.029 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.029 [ 00:09:52.029 { 00:09:52.029 "name": "NewBaseBdev", 00:09:52.029 "aliases": [ 00:09:52.029 "5e24f951-48ce-4a67-90a0-0fe4c3cf03d1" 00:09:52.029 ], 00:09:52.029 "product_name": "Malloc disk", 00:09:52.029 "block_size": 512, 00:09:52.029 "num_blocks": 65536, 00:09:52.029 "uuid": "5e24f951-48ce-4a67-90a0-0fe4c3cf03d1", 00:09:52.029 "assigned_rate_limits": { 00:09:52.029 "rw_ios_per_sec": 0, 00:09:52.029 "rw_mbytes_per_sec": 0, 00:09:52.029 "r_mbytes_per_sec": 0, 00:09:52.029 "w_mbytes_per_sec": 0 00:09:52.029 }, 00:09:52.029 "claimed": true, 00:09:52.029 "claim_type": "exclusive_write", 00:09:52.029 "zoned": false, 00:09:52.029 "supported_io_types": { 00:09:52.029 "read": true, 00:09:52.029 "write": true, 00:09:52.029 "unmap": true, 00:09:52.029 "flush": true, 00:09:52.029 "reset": true, 00:09:52.029 "nvme_admin": false, 00:09:52.029 "nvme_io": false, 00:09:52.029 "nvme_io_md": false, 00:09:52.029 "write_zeroes": true, 00:09:52.029 "zcopy": true, 00:09:52.029 "get_zone_info": false, 00:09:52.029 "zone_management": false, 00:09:52.029 "zone_append": false, 00:09:52.029 "compare": false, 00:09:52.029 "compare_and_write": false, 00:09:52.029 "abort": true, 00:09:52.029 "seek_hole": false, 00:09:52.029 "seek_data": false, 00:09:52.029 "copy": true, 00:09:52.029 "nvme_iov_md": false 00:09:52.029 }, 00:09:52.029 "memory_domains": [ 00:09:52.029 { 00:09:52.029 "dma_device_id": "system", 00:09:52.029 "dma_device_type": 1 00:09:52.029 }, 00:09:52.029 { 00:09:52.029 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:52.029 "dma_device_type": 2 00:09:52.029 } 00:09:52.029 ], 00:09:52.029 "driver_specific": {} 00:09:52.029 } 00:09:52.029 ] 00:09:52.029 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.029 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:52.029 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:09:52.029 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:52.029 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:52.029 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:52.029 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:52.029 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:52.029 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:52.029 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:52.029 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:52.029 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:52.029 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:52.029 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:52.029 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.029 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.029 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.029 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:52.029 "name": "Existed_Raid", 00:09:52.029 "uuid": "1531ef53-6991-441e-a935-5e20321565d7", 00:09:52.029 "strip_size_kb": 64, 00:09:52.029 "state": "online", 00:09:52.029 "raid_level": "concat", 00:09:52.029 "superblock": false, 00:09:52.029 "num_base_bdevs": 3, 00:09:52.029 "num_base_bdevs_discovered": 3, 00:09:52.029 "num_base_bdevs_operational": 3, 00:09:52.029 "base_bdevs_list": [ 00:09:52.029 { 00:09:52.029 "name": "NewBaseBdev", 00:09:52.029 "uuid": "5e24f951-48ce-4a67-90a0-0fe4c3cf03d1", 00:09:52.029 "is_configured": true, 00:09:52.029 "data_offset": 0, 00:09:52.029 "data_size": 65536 00:09:52.029 }, 00:09:52.029 { 00:09:52.029 "name": "BaseBdev2", 00:09:52.029 "uuid": "ede0b109-be63-4ba3-bd3f-40824b74fd9e", 00:09:52.029 "is_configured": true, 00:09:52.029 "data_offset": 0, 00:09:52.029 "data_size": 65536 00:09:52.029 }, 00:09:52.029 { 00:09:52.029 "name": "BaseBdev3", 00:09:52.029 "uuid": "aa2de657-01e9-473e-998c-19cae8d88e03", 00:09:52.029 "is_configured": true, 00:09:52.029 "data_offset": 0, 00:09:52.029 "data_size": 65536 00:09:52.029 } 00:09:52.029 ] 00:09:52.029 }' 00:09:52.029 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:52.029 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.599 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:09:52.599 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:52.599 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:52.599 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:52.599 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:52.599 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:52.599 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:52.599 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.599 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.599 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:52.599 [2024-11-19 15:15:42.702503] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:52.599 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.599 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:52.599 "name": "Existed_Raid", 00:09:52.599 "aliases": [ 00:09:52.599 "1531ef53-6991-441e-a935-5e20321565d7" 00:09:52.599 ], 00:09:52.599 "product_name": "Raid Volume", 00:09:52.599 "block_size": 512, 00:09:52.599 "num_blocks": 196608, 00:09:52.599 "uuid": "1531ef53-6991-441e-a935-5e20321565d7", 00:09:52.599 "assigned_rate_limits": { 00:09:52.599 "rw_ios_per_sec": 0, 00:09:52.599 "rw_mbytes_per_sec": 0, 00:09:52.599 "r_mbytes_per_sec": 0, 00:09:52.599 "w_mbytes_per_sec": 0 00:09:52.599 }, 00:09:52.599 "claimed": false, 00:09:52.599 "zoned": false, 00:09:52.599 "supported_io_types": { 00:09:52.599 "read": true, 00:09:52.599 "write": true, 00:09:52.599 "unmap": true, 00:09:52.599 "flush": true, 00:09:52.599 "reset": true, 00:09:52.599 "nvme_admin": false, 00:09:52.599 "nvme_io": false, 00:09:52.599 "nvme_io_md": false, 00:09:52.599 "write_zeroes": true, 00:09:52.599 "zcopy": false, 00:09:52.599 "get_zone_info": false, 00:09:52.599 "zone_management": false, 00:09:52.599 "zone_append": false, 00:09:52.599 "compare": false, 00:09:52.599 "compare_and_write": false, 00:09:52.599 "abort": false, 00:09:52.599 "seek_hole": false, 00:09:52.599 "seek_data": false, 00:09:52.599 "copy": false, 00:09:52.599 "nvme_iov_md": false 00:09:52.599 }, 00:09:52.599 "memory_domains": [ 00:09:52.599 { 00:09:52.599 "dma_device_id": "system", 00:09:52.599 "dma_device_type": 1 00:09:52.599 }, 00:09:52.599 { 00:09:52.599 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:52.599 "dma_device_type": 2 00:09:52.599 }, 00:09:52.599 { 00:09:52.599 "dma_device_id": "system", 00:09:52.599 "dma_device_type": 1 00:09:52.599 }, 00:09:52.599 { 00:09:52.599 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:52.599 "dma_device_type": 2 00:09:52.599 }, 00:09:52.599 { 00:09:52.599 "dma_device_id": "system", 00:09:52.599 "dma_device_type": 1 00:09:52.599 }, 00:09:52.599 { 00:09:52.599 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:52.599 "dma_device_type": 2 00:09:52.599 } 00:09:52.599 ], 00:09:52.599 "driver_specific": { 00:09:52.599 "raid": { 00:09:52.599 "uuid": "1531ef53-6991-441e-a935-5e20321565d7", 00:09:52.599 "strip_size_kb": 64, 00:09:52.599 "state": "online", 00:09:52.599 "raid_level": "concat", 00:09:52.599 "superblock": false, 00:09:52.599 "num_base_bdevs": 3, 00:09:52.599 "num_base_bdevs_discovered": 3, 00:09:52.599 "num_base_bdevs_operational": 3, 00:09:52.599 "base_bdevs_list": [ 00:09:52.599 { 00:09:52.599 "name": "NewBaseBdev", 00:09:52.599 "uuid": "5e24f951-48ce-4a67-90a0-0fe4c3cf03d1", 00:09:52.599 "is_configured": true, 00:09:52.599 "data_offset": 0, 00:09:52.599 "data_size": 65536 00:09:52.599 }, 00:09:52.599 { 00:09:52.599 "name": "BaseBdev2", 00:09:52.599 "uuid": "ede0b109-be63-4ba3-bd3f-40824b74fd9e", 00:09:52.599 "is_configured": true, 00:09:52.599 "data_offset": 0, 00:09:52.599 "data_size": 65536 00:09:52.599 }, 00:09:52.599 { 00:09:52.599 "name": "BaseBdev3", 00:09:52.599 "uuid": "aa2de657-01e9-473e-998c-19cae8d88e03", 00:09:52.599 "is_configured": true, 00:09:52.599 "data_offset": 0, 00:09:52.599 "data_size": 65536 00:09:52.599 } 00:09:52.599 ] 00:09:52.599 } 00:09:52.599 } 00:09:52.599 }' 00:09:52.599 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:52.599 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:09:52.599 BaseBdev2 00:09:52.599 BaseBdev3' 00:09:52.599 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:52.599 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:52.599 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:52.599 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:52.599 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:09:52.599 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.599 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.599 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.599 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:52.599 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:52.599 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:52.599 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:52.599 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.599 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.599 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:52.599 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.599 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:52.599 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:52.599 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:52.600 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:52.600 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:52.600 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.600 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.600 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.859 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:52.859 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:52.859 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:52.859 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.859 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.859 [2024-11-19 15:15:42.961738] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:52.859 [2024-11-19 15:15:42.961853] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:52.859 [2024-11-19 15:15:42.961974] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:52.859 [2024-11-19 15:15:42.962065] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:52.859 [2024-11-19 15:15:42.962093] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:09:52.859 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.859 15:15:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 76742 00:09:52.859 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 76742 ']' 00:09:52.859 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 76742 00:09:52.859 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:09:52.859 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:52.859 15:15:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 76742 00:09:52.859 15:15:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:52.859 15:15:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:52.859 15:15:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 76742' 00:09:52.859 killing process with pid 76742 00:09:52.859 15:15:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 76742 00:09:52.859 [2024-11-19 15:15:43.009501] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:52.859 15:15:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 76742 00:09:52.859 [2024-11-19 15:15:43.069480] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:53.117 15:15:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:09:53.117 00:09:53.117 real 0m8.815s 00:09:53.117 user 0m14.781s 00:09:53.117 sys 0m1.929s 00:09:53.117 15:15:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:53.117 15:15:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:53.117 ************************************ 00:09:53.117 END TEST raid_state_function_test 00:09:53.117 ************************************ 00:09:53.117 15:15:43 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test concat 3 true 00:09:53.376 15:15:43 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:53.376 15:15:43 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:53.376 15:15:43 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:53.376 ************************************ 00:09:53.376 START TEST raid_state_function_test_sb 00:09:53.376 ************************************ 00:09:53.376 15:15:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test concat 3 true 00:09:53.376 15:15:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:09:53.376 15:15:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:09:53.376 15:15:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:09:53.376 15:15:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:09:53.376 15:15:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:09:53.376 15:15:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:53.376 15:15:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:09:53.376 15:15:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:53.376 15:15:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:53.376 15:15:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:09:53.376 15:15:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:53.376 15:15:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:53.376 15:15:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:09:53.376 15:15:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:53.376 15:15:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:53.376 15:15:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:09:53.376 15:15:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:09:53.376 15:15:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:09:53.376 15:15:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:09:53.376 15:15:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:09:53.376 15:15:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:09:53.377 15:15:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:09:53.377 15:15:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:09:53.377 15:15:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:09:53.377 15:15:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:09:53.377 15:15:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:09:53.377 Process raid pid: 77341 00:09:53.377 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:53.377 15:15:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=77341 00:09:53.377 15:15:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:09:53.377 15:15:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 77341' 00:09:53.377 15:15:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 77341 00:09:53.377 15:15:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 77341 ']' 00:09:53.377 15:15:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:53.377 15:15:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:53.377 15:15:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:53.377 15:15:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:53.377 15:15:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:53.377 [2024-11-19 15:15:43.560996] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:09:53.377 [2024-11-19 15:15:43.561201] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:53.377 [2024-11-19 15:15:43.696461] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:53.636 [2024-11-19 15:15:43.740189] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:53.636 [2024-11-19 15:15:43.816896] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:53.636 [2024-11-19 15:15:43.817075] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:54.205 15:15:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:54.205 15:15:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:09:54.205 15:15:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:54.205 15:15:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:54.205 15:15:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:54.205 [2024-11-19 15:15:44.424518] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:54.205 [2024-11-19 15:15:44.424701] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:54.205 [2024-11-19 15:15:44.424740] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:54.205 [2024-11-19 15:15:44.424771] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:54.205 [2024-11-19 15:15:44.424793] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:54.205 [2024-11-19 15:15:44.424821] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:54.205 15:15:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.205 15:15:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:54.205 15:15:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:54.205 15:15:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:54.205 15:15:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:54.205 15:15:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:54.205 15:15:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:54.205 15:15:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:54.205 15:15:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:54.205 15:15:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:54.205 15:15:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:54.205 15:15:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:54.205 15:15:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:54.205 15:15:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:54.205 15:15:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:54.205 15:15:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.205 15:15:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:54.205 "name": "Existed_Raid", 00:09:54.205 "uuid": "d08e5ae9-c6cc-43ef-a586-e603a7b4d0a3", 00:09:54.205 "strip_size_kb": 64, 00:09:54.206 "state": "configuring", 00:09:54.206 "raid_level": "concat", 00:09:54.206 "superblock": true, 00:09:54.206 "num_base_bdevs": 3, 00:09:54.206 "num_base_bdevs_discovered": 0, 00:09:54.206 "num_base_bdevs_operational": 3, 00:09:54.206 "base_bdevs_list": [ 00:09:54.206 { 00:09:54.206 "name": "BaseBdev1", 00:09:54.206 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:54.206 "is_configured": false, 00:09:54.206 "data_offset": 0, 00:09:54.206 "data_size": 0 00:09:54.206 }, 00:09:54.206 { 00:09:54.206 "name": "BaseBdev2", 00:09:54.206 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:54.206 "is_configured": false, 00:09:54.206 "data_offset": 0, 00:09:54.206 "data_size": 0 00:09:54.206 }, 00:09:54.206 { 00:09:54.206 "name": "BaseBdev3", 00:09:54.206 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:54.206 "is_configured": false, 00:09:54.206 "data_offset": 0, 00:09:54.206 "data_size": 0 00:09:54.206 } 00:09:54.206 ] 00:09:54.206 }' 00:09:54.206 15:15:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:54.206 15:15:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:54.774 [2024-11-19 15:15:44.843698] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:54.774 [2024-11-19 15:15:44.843790] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:54.774 [2024-11-19 15:15:44.855754] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:54.774 [2024-11-19 15:15:44.855916] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:54.774 [2024-11-19 15:15:44.855930] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:54.774 [2024-11-19 15:15:44.855941] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:54.774 [2024-11-19 15:15:44.855949] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:54.774 [2024-11-19 15:15:44.855959] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:54.774 [2024-11-19 15:15:44.883124] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:54.774 BaseBdev1 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:54.774 [ 00:09:54.774 { 00:09:54.774 "name": "BaseBdev1", 00:09:54.774 "aliases": [ 00:09:54.774 "a3af1df0-9a39-4d4c-b7e0-59f901b7d265" 00:09:54.774 ], 00:09:54.774 "product_name": "Malloc disk", 00:09:54.774 "block_size": 512, 00:09:54.774 "num_blocks": 65536, 00:09:54.774 "uuid": "a3af1df0-9a39-4d4c-b7e0-59f901b7d265", 00:09:54.774 "assigned_rate_limits": { 00:09:54.774 "rw_ios_per_sec": 0, 00:09:54.774 "rw_mbytes_per_sec": 0, 00:09:54.774 "r_mbytes_per_sec": 0, 00:09:54.774 "w_mbytes_per_sec": 0 00:09:54.774 }, 00:09:54.774 "claimed": true, 00:09:54.774 "claim_type": "exclusive_write", 00:09:54.774 "zoned": false, 00:09:54.774 "supported_io_types": { 00:09:54.774 "read": true, 00:09:54.774 "write": true, 00:09:54.774 "unmap": true, 00:09:54.774 "flush": true, 00:09:54.774 "reset": true, 00:09:54.774 "nvme_admin": false, 00:09:54.774 "nvme_io": false, 00:09:54.774 "nvme_io_md": false, 00:09:54.774 "write_zeroes": true, 00:09:54.774 "zcopy": true, 00:09:54.774 "get_zone_info": false, 00:09:54.774 "zone_management": false, 00:09:54.774 "zone_append": false, 00:09:54.774 "compare": false, 00:09:54.774 "compare_and_write": false, 00:09:54.774 "abort": true, 00:09:54.774 "seek_hole": false, 00:09:54.774 "seek_data": false, 00:09:54.774 "copy": true, 00:09:54.774 "nvme_iov_md": false 00:09:54.774 }, 00:09:54.774 "memory_domains": [ 00:09:54.774 { 00:09:54.774 "dma_device_id": "system", 00:09:54.774 "dma_device_type": 1 00:09:54.774 }, 00:09:54.774 { 00:09:54.774 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:54.774 "dma_device_type": 2 00:09:54.774 } 00:09:54.774 ], 00:09:54.774 "driver_specific": {} 00:09:54.774 } 00:09:54.774 ] 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:54.774 "name": "Existed_Raid", 00:09:54.774 "uuid": "28408aca-d7d5-45fc-a227-38a945e80a5c", 00:09:54.774 "strip_size_kb": 64, 00:09:54.774 "state": "configuring", 00:09:54.774 "raid_level": "concat", 00:09:54.774 "superblock": true, 00:09:54.774 "num_base_bdevs": 3, 00:09:54.774 "num_base_bdevs_discovered": 1, 00:09:54.774 "num_base_bdevs_operational": 3, 00:09:54.774 "base_bdevs_list": [ 00:09:54.774 { 00:09:54.774 "name": "BaseBdev1", 00:09:54.774 "uuid": "a3af1df0-9a39-4d4c-b7e0-59f901b7d265", 00:09:54.774 "is_configured": true, 00:09:54.774 "data_offset": 2048, 00:09:54.774 "data_size": 63488 00:09:54.774 }, 00:09:54.774 { 00:09:54.774 "name": "BaseBdev2", 00:09:54.774 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:54.774 "is_configured": false, 00:09:54.774 "data_offset": 0, 00:09:54.774 "data_size": 0 00:09:54.774 }, 00:09:54.774 { 00:09:54.774 "name": "BaseBdev3", 00:09:54.774 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:54.774 "is_configured": false, 00:09:54.774 "data_offset": 0, 00:09:54.774 "data_size": 0 00:09:54.774 } 00:09:54.774 ] 00:09:54.774 }' 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:54.774 15:15:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:55.343 15:15:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:55.343 15:15:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:55.343 15:15:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:55.343 [2024-11-19 15:15:45.394322] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:55.343 [2024-11-19 15:15:45.394483] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:09:55.343 15:15:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:55.343 15:15:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:55.343 15:15:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:55.343 15:15:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:55.343 [2024-11-19 15:15:45.402334] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:55.343 [2024-11-19 15:15:45.404491] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:55.343 [2024-11-19 15:15:45.404537] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:55.343 [2024-11-19 15:15:45.404547] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:55.343 [2024-11-19 15:15:45.404559] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:55.343 15:15:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:55.343 15:15:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:09:55.343 15:15:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:55.343 15:15:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:55.343 15:15:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:55.343 15:15:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:55.343 15:15:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:55.343 15:15:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:55.343 15:15:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:55.343 15:15:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:55.343 15:15:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:55.343 15:15:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:55.343 15:15:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:55.343 15:15:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:55.343 15:15:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:55.343 15:15:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:55.343 15:15:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:55.343 15:15:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:55.343 15:15:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:55.343 "name": "Existed_Raid", 00:09:55.343 "uuid": "630c9f58-f3bd-4014-b787-b215bd4dbd9c", 00:09:55.343 "strip_size_kb": 64, 00:09:55.343 "state": "configuring", 00:09:55.343 "raid_level": "concat", 00:09:55.343 "superblock": true, 00:09:55.343 "num_base_bdevs": 3, 00:09:55.343 "num_base_bdevs_discovered": 1, 00:09:55.343 "num_base_bdevs_operational": 3, 00:09:55.343 "base_bdevs_list": [ 00:09:55.343 { 00:09:55.343 "name": "BaseBdev1", 00:09:55.343 "uuid": "a3af1df0-9a39-4d4c-b7e0-59f901b7d265", 00:09:55.343 "is_configured": true, 00:09:55.343 "data_offset": 2048, 00:09:55.343 "data_size": 63488 00:09:55.343 }, 00:09:55.343 { 00:09:55.343 "name": "BaseBdev2", 00:09:55.343 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:55.343 "is_configured": false, 00:09:55.343 "data_offset": 0, 00:09:55.343 "data_size": 0 00:09:55.343 }, 00:09:55.343 { 00:09:55.343 "name": "BaseBdev3", 00:09:55.343 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:55.343 "is_configured": false, 00:09:55.343 "data_offset": 0, 00:09:55.343 "data_size": 0 00:09:55.343 } 00:09:55.343 ] 00:09:55.343 }' 00:09:55.343 15:15:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:55.343 15:15:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:55.604 15:15:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:55.604 15:15:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:55.604 15:15:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:55.604 [2024-11-19 15:15:45.830312] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:55.604 BaseBdev2 00:09:55.604 15:15:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:55.604 15:15:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:09:55.604 15:15:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:55.604 15:15:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:55.604 15:15:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:55.604 15:15:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:55.604 15:15:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:55.604 15:15:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:55.604 15:15:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:55.604 15:15:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:55.604 15:15:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:55.604 15:15:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:55.604 15:15:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:55.604 15:15:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:55.604 [ 00:09:55.604 { 00:09:55.604 "name": "BaseBdev2", 00:09:55.604 "aliases": [ 00:09:55.604 "8892f81d-f703-47ab-983d-df0b6da9785f" 00:09:55.604 ], 00:09:55.604 "product_name": "Malloc disk", 00:09:55.604 "block_size": 512, 00:09:55.604 "num_blocks": 65536, 00:09:55.604 "uuid": "8892f81d-f703-47ab-983d-df0b6da9785f", 00:09:55.604 "assigned_rate_limits": { 00:09:55.604 "rw_ios_per_sec": 0, 00:09:55.604 "rw_mbytes_per_sec": 0, 00:09:55.604 "r_mbytes_per_sec": 0, 00:09:55.604 "w_mbytes_per_sec": 0 00:09:55.604 }, 00:09:55.604 "claimed": true, 00:09:55.604 "claim_type": "exclusive_write", 00:09:55.604 "zoned": false, 00:09:55.604 "supported_io_types": { 00:09:55.604 "read": true, 00:09:55.604 "write": true, 00:09:55.604 "unmap": true, 00:09:55.604 "flush": true, 00:09:55.604 "reset": true, 00:09:55.604 "nvme_admin": false, 00:09:55.604 "nvme_io": false, 00:09:55.604 "nvme_io_md": false, 00:09:55.604 "write_zeroes": true, 00:09:55.604 "zcopy": true, 00:09:55.604 "get_zone_info": false, 00:09:55.604 "zone_management": false, 00:09:55.604 "zone_append": false, 00:09:55.604 "compare": false, 00:09:55.604 "compare_and_write": false, 00:09:55.604 "abort": true, 00:09:55.604 "seek_hole": false, 00:09:55.604 "seek_data": false, 00:09:55.604 "copy": true, 00:09:55.604 "nvme_iov_md": false 00:09:55.604 }, 00:09:55.604 "memory_domains": [ 00:09:55.604 { 00:09:55.604 "dma_device_id": "system", 00:09:55.604 "dma_device_type": 1 00:09:55.604 }, 00:09:55.604 { 00:09:55.604 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:55.604 "dma_device_type": 2 00:09:55.604 } 00:09:55.604 ], 00:09:55.604 "driver_specific": {} 00:09:55.604 } 00:09:55.604 ] 00:09:55.604 15:15:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:55.604 15:15:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:55.604 15:15:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:55.604 15:15:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:55.604 15:15:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:55.604 15:15:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:55.604 15:15:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:55.604 15:15:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:55.604 15:15:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:55.604 15:15:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:55.604 15:15:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:55.604 15:15:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:55.604 15:15:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:55.604 15:15:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:55.604 15:15:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:55.604 15:15:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:55.604 15:15:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:55.604 15:15:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:55.604 15:15:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:55.604 15:15:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:55.604 "name": "Existed_Raid", 00:09:55.604 "uuid": "630c9f58-f3bd-4014-b787-b215bd4dbd9c", 00:09:55.604 "strip_size_kb": 64, 00:09:55.604 "state": "configuring", 00:09:55.604 "raid_level": "concat", 00:09:55.604 "superblock": true, 00:09:55.604 "num_base_bdevs": 3, 00:09:55.604 "num_base_bdevs_discovered": 2, 00:09:55.604 "num_base_bdevs_operational": 3, 00:09:55.604 "base_bdevs_list": [ 00:09:55.604 { 00:09:55.604 "name": "BaseBdev1", 00:09:55.604 "uuid": "a3af1df0-9a39-4d4c-b7e0-59f901b7d265", 00:09:55.604 "is_configured": true, 00:09:55.604 "data_offset": 2048, 00:09:55.604 "data_size": 63488 00:09:55.604 }, 00:09:55.604 { 00:09:55.604 "name": "BaseBdev2", 00:09:55.604 "uuid": "8892f81d-f703-47ab-983d-df0b6da9785f", 00:09:55.604 "is_configured": true, 00:09:55.604 "data_offset": 2048, 00:09:55.604 "data_size": 63488 00:09:55.604 }, 00:09:55.604 { 00:09:55.604 "name": "BaseBdev3", 00:09:55.604 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:55.604 "is_configured": false, 00:09:55.604 "data_offset": 0, 00:09:55.604 "data_size": 0 00:09:55.604 } 00:09:55.604 ] 00:09:55.604 }' 00:09:55.604 15:15:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:55.604 15:15:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:56.174 15:15:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:56.174 15:15:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:56.174 15:15:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:56.174 [2024-11-19 15:15:46.336755] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:56.174 [2024-11-19 15:15:46.337053] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:09:56.174 [2024-11-19 15:15:46.337082] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:09:56.174 BaseBdev3 00:09:56.174 [2024-11-19 15:15:46.337517] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:09:56.174 [2024-11-19 15:15:46.337720] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:09:56.174 [2024-11-19 15:15:46.337736] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:09:56.174 [2024-11-19 15:15:46.337916] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:56.174 15:15:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:56.174 15:15:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:09:56.174 15:15:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:09:56.174 15:15:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:56.174 15:15:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:56.174 15:15:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:56.174 15:15:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:56.174 15:15:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:56.174 15:15:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:56.174 15:15:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:56.174 15:15:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:56.174 15:15:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:56.174 15:15:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:56.174 15:15:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:56.174 [ 00:09:56.174 { 00:09:56.174 "name": "BaseBdev3", 00:09:56.174 "aliases": [ 00:09:56.174 "cf780bae-cd7c-45d4-b816-ba336651e38e" 00:09:56.174 ], 00:09:56.174 "product_name": "Malloc disk", 00:09:56.174 "block_size": 512, 00:09:56.174 "num_blocks": 65536, 00:09:56.174 "uuid": "cf780bae-cd7c-45d4-b816-ba336651e38e", 00:09:56.174 "assigned_rate_limits": { 00:09:56.174 "rw_ios_per_sec": 0, 00:09:56.174 "rw_mbytes_per_sec": 0, 00:09:56.174 "r_mbytes_per_sec": 0, 00:09:56.174 "w_mbytes_per_sec": 0 00:09:56.174 }, 00:09:56.175 "claimed": true, 00:09:56.175 "claim_type": "exclusive_write", 00:09:56.175 "zoned": false, 00:09:56.175 "supported_io_types": { 00:09:56.175 "read": true, 00:09:56.175 "write": true, 00:09:56.175 "unmap": true, 00:09:56.175 "flush": true, 00:09:56.175 "reset": true, 00:09:56.175 "nvme_admin": false, 00:09:56.175 "nvme_io": false, 00:09:56.175 "nvme_io_md": false, 00:09:56.175 "write_zeroes": true, 00:09:56.175 "zcopy": true, 00:09:56.175 "get_zone_info": false, 00:09:56.175 "zone_management": false, 00:09:56.175 "zone_append": false, 00:09:56.175 "compare": false, 00:09:56.175 "compare_and_write": false, 00:09:56.175 "abort": true, 00:09:56.175 "seek_hole": false, 00:09:56.175 "seek_data": false, 00:09:56.175 "copy": true, 00:09:56.175 "nvme_iov_md": false 00:09:56.175 }, 00:09:56.175 "memory_domains": [ 00:09:56.175 { 00:09:56.175 "dma_device_id": "system", 00:09:56.175 "dma_device_type": 1 00:09:56.175 }, 00:09:56.175 { 00:09:56.175 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:56.175 "dma_device_type": 2 00:09:56.175 } 00:09:56.175 ], 00:09:56.175 "driver_specific": {} 00:09:56.175 } 00:09:56.175 ] 00:09:56.175 15:15:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:56.175 15:15:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:56.175 15:15:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:56.175 15:15:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:56.175 15:15:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:09:56.175 15:15:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:56.175 15:15:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:56.175 15:15:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:56.175 15:15:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:56.175 15:15:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:56.175 15:15:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:56.175 15:15:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:56.175 15:15:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:56.175 15:15:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:56.175 15:15:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:56.175 15:15:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:56.175 15:15:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:56.175 15:15:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:56.175 15:15:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:56.175 15:15:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:56.175 "name": "Existed_Raid", 00:09:56.175 "uuid": "630c9f58-f3bd-4014-b787-b215bd4dbd9c", 00:09:56.175 "strip_size_kb": 64, 00:09:56.175 "state": "online", 00:09:56.175 "raid_level": "concat", 00:09:56.175 "superblock": true, 00:09:56.175 "num_base_bdevs": 3, 00:09:56.175 "num_base_bdevs_discovered": 3, 00:09:56.175 "num_base_bdevs_operational": 3, 00:09:56.175 "base_bdevs_list": [ 00:09:56.175 { 00:09:56.175 "name": "BaseBdev1", 00:09:56.175 "uuid": "a3af1df0-9a39-4d4c-b7e0-59f901b7d265", 00:09:56.175 "is_configured": true, 00:09:56.175 "data_offset": 2048, 00:09:56.175 "data_size": 63488 00:09:56.175 }, 00:09:56.175 { 00:09:56.175 "name": "BaseBdev2", 00:09:56.175 "uuid": "8892f81d-f703-47ab-983d-df0b6da9785f", 00:09:56.175 "is_configured": true, 00:09:56.175 "data_offset": 2048, 00:09:56.175 "data_size": 63488 00:09:56.175 }, 00:09:56.175 { 00:09:56.175 "name": "BaseBdev3", 00:09:56.175 "uuid": "cf780bae-cd7c-45d4-b816-ba336651e38e", 00:09:56.175 "is_configured": true, 00:09:56.175 "data_offset": 2048, 00:09:56.175 "data_size": 63488 00:09:56.175 } 00:09:56.175 ] 00:09:56.175 }' 00:09:56.175 15:15:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:56.175 15:15:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:56.745 15:15:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:09:56.745 15:15:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:56.745 15:15:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:56.745 15:15:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:56.745 15:15:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:09:56.745 15:15:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:56.745 15:15:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:56.745 15:15:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:56.745 15:15:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:56.745 15:15:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:56.745 [2024-11-19 15:15:46.828435] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:56.745 15:15:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:56.745 15:15:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:56.745 "name": "Existed_Raid", 00:09:56.745 "aliases": [ 00:09:56.745 "630c9f58-f3bd-4014-b787-b215bd4dbd9c" 00:09:56.745 ], 00:09:56.745 "product_name": "Raid Volume", 00:09:56.745 "block_size": 512, 00:09:56.745 "num_blocks": 190464, 00:09:56.745 "uuid": "630c9f58-f3bd-4014-b787-b215bd4dbd9c", 00:09:56.745 "assigned_rate_limits": { 00:09:56.745 "rw_ios_per_sec": 0, 00:09:56.745 "rw_mbytes_per_sec": 0, 00:09:56.745 "r_mbytes_per_sec": 0, 00:09:56.745 "w_mbytes_per_sec": 0 00:09:56.745 }, 00:09:56.745 "claimed": false, 00:09:56.745 "zoned": false, 00:09:56.745 "supported_io_types": { 00:09:56.745 "read": true, 00:09:56.745 "write": true, 00:09:56.745 "unmap": true, 00:09:56.745 "flush": true, 00:09:56.745 "reset": true, 00:09:56.745 "nvme_admin": false, 00:09:56.745 "nvme_io": false, 00:09:56.745 "nvme_io_md": false, 00:09:56.745 "write_zeroes": true, 00:09:56.745 "zcopy": false, 00:09:56.745 "get_zone_info": false, 00:09:56.745 "zone_management": false, 00:09:56.745 "zone_append": false, 00:09:56.745 "compare": false, 00:09:56.745 "compare_and_write": false, 00:09:56.745 "abort": false, 00:09:56.745 "seek_hole": false, 00:09:56.745 "seek_data": false, 00:09:56.745 "copy": false, 00:09:56.745 "nvme_iov_md": false 00:09:56.745 }, 00:09:56.745 "memory_domains": [ 00:09:56.745 { 00:09:56.745 "dma_device_id": "system", 00:09:56.745 "dma_device_type": 1 00:09:56.745 }, 00:09:56.745 { 00:09:56.745 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:56.745 "dma_device_type": 2 00:09:56.745 }, 00:09:56.745 { 00:09:56.745 "dma_device_id": "system", 00:09:56.745 "dma_device_type": 1 00:09:56.745 }, 00:09:56.745 { 00:09:56.745 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:56.745 "dma_device_type": 2 00:09:56.745 }, 00:09:56.745 { 00:09:56.745 "dma_device_id": "system", 00:09:56.745 "dma_device_type": 1 00:09:56.745 }, 00:09:56.745 { 00:09:56.745 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:56.745 "dma_device_type": 2 00:09:56.745 } 00:09:56.745 ], 00:09:56.745 "driver_specific": { 00:09:56.745 "raid": { 00:09:56.745 "uuid": "630c9f58-f3bd-4014-b787-b215bd4dbd9c", 00:09:56.745 "strip_size_kb": 64, 00:09:56.745 "state": "online", 00:09:56.745 "raid_level": "concat", 00:09:56.745 "superblock": true, 00:09:56.745 "num_base_bdevs": 3, 00:09:56.745 "num_base_bdevs_discovered": 3, 00:09:56.745 "num_base_bdevs_operational": 3, 00:09:56.745 "base_bdevs_list": [ 00:09:56.745 { 00:09:56.745 "name": "BaseBdev1", 00:09:56.745 "uuid": "a3af1df0-9a39-4d4c-b7e0-59f901b7d265", 00:09:56.745 "is_configured": true, 00:09:56.745 "data_offset": 2048, 00:09:56.745 "data_size": 63488 00:09:56.745 }, 00:09:56.745 { 00:09:56.745 "name": "BaseBdev2", 00:09:56.745 "uuid": "8892f81d-f703-47ab-983d-df0b6da9785f", 00:09:56.745 "is_configured": true, 00:09:56.745 "data_offset": 2048, 00:09:56.745 "data_size": 63488 00:09:56.745 }, 00:09:56.745 { 00:09:56.745 "name": "BaseBdev3", 00:09:56.745 "uuid": "cf780bae-cd7c-45d4-b816-ba336651e38e", 00:09:56.745 "is_configured": true, 00:09:56.745 "data_offset": 2048, 00:09:56.745 "data_size": 63488 00:09:56.745 } 00:09:56.745 ] 00:09:56.745 } 00:09:56.745 } 00:09:56.745 }' 00:09:56.745 15:15:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:56.745 15:15:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:09:56.745 BaseBdev2 00:09:56.745 BaseBdev3' 00:09:56.745 15:15:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:56.745 15:15:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:56.745 15:15:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:56.745 15:15:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:09:56.745 15:15:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:56.745 15:15:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:56.745 15:15:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:56.745 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:56.745 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:56.745 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:56.745 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:56.745 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:56.745 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:56.745 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:56.745 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:56.745 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.006 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:57.006 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:57.006 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:57.006 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:57.006 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:57.006 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.006 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.006 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.006 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:57.006 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:57.006 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:57.006 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.006 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.006 [2024-11-19 15:15:47.119919] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:57.006 [2024-11-19 15:15:47.119985] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:57.006 [2024-11-19 15:15:47.120072] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:57.006 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.006 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:09:57.006 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:09:57.006 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:57.006 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:09:57.006 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:09:57.006 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 2 00:09:57.006 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:57.006 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:09:57.006 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:57.006 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:57.006 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:57.006 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:57.006 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:57.006 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:57.006 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:57.006 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:57.006 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.006 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.006 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:57.006 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.006 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:57.006 "name": "Existed_Raid", 00:09:57.006 "uuid": "630c9f58-f3bd-4014-b787-b215bd4dbd9c", 00:09:57.006 "strip_size_kb": 64, 00:09:57.006 "state": "offline", 00:09:57.006 "raid_level": "concat", 00:09:57.006 "superblock": true, 00:09:57.006 "num_base_bdevs": 3, 00:09:57.006 "num_base_bdevs_discovered": 2, 00:09:57.006 "num_base_bdevs_operational": 2, 00:09:57.006 "base_bdevs_list": [ 00:09:57.006 { 00:09:57.006 "name": null, 00:09:57.006 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:57.006 "is_configured": false, 00:09:57.006 "data_offset": 0, 00:09:57.006 "data_size": 63488 00:09:57.007 }, 00:09:57.007 { 00:09:57.007 "name": "BaseBdev2", 00:09:57.007 "uuid": "8892f81d-f703-47ab-983d-df0b6da9785f", 00:09:57.007 "is_configured": true, 00:09:57.007 "data_offset": 2048, 00:09:57.007 "data_size": 63488 00:09:57.007 }, 00:09:57.007 { 00:09:57.007 "name": "BaseBdev3", 00:09:57.007 "uuid": "cf780bae-cd7c-45d4-b816-ba336651e38e", 00:09:57.007 "is_configured": true, 00:09:57.007 "data_offset": 2048, 00:09:57.007 "data_size": 63488 00:09:57.007 } 00:09:57.007 ] 00:09:57.007 }' 00:09:57.007 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:57.007 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.267 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:09:57.267 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:57.527 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:57.527 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.527 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.527 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:57.527 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.527 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:57.527 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:57.527 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:09:57.527 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.527 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.527 [2024-11-19 15:15:47.640414] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:57.527 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.527 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:57.527 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:57.527 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:57.527 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.527 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.527 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:57.527 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.528 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:57.528 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:57.528 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:09:57.528 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.528 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.528 [2024-11-19 15:15:47.717244] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:57.528 [2024-11-19 15:15:47.717325] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:09:57.528 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.528 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:57.528 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:57.528 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:57.528 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.528 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:09:57.528 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.528 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.528 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:09:57.528 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:09:57.528 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:09:57.528 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:09:57.528 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:57.528 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:57.528 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.528 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.528 BaseBdev2 00:09:57.528 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.528 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:09:57.528 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:57.528 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:57.528 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:57.528 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:57.528 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:57.528 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:57.528 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.528 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.528 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.528 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:57.528 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.528 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.528 [ 00:09:57.528 { 00:09:57.528 "name": "BaseBdev2", 00:09:57.528 "aliases": [ 00:09:57.528 "6dff4074-657f-49cb-8046-a00c33ef1c1c" 00:09:57.528 ], 00:09:57.528 "product_name": "Malloc disk", 00:09:57.528 "block_size": 512, 00:09:57.528 "num_blocks": 65536, 00:09:57.528 "uuid": "6dff4074-657f-49cb-8046-a00c33ef1c1c", 00:09:57.528 "assigned_rate_limits": { 00:09:57.528 "rw_ios_per_sec": 0, 00:09:57.528 "rw_mbytes_per_sec": 0, 00:09:57.528 "r_mbytes_per_sec": 0, 00:09:57.528 "w_mbytes_per_sec": 0 00:09:57.528 }, 00:09:57.528 "claimed": false, 00:09:57.528 "zoned": false, 00:09:57.528 "supported_io_types": { 00:09:57.528 "read": true, 00:09:57.528 "write": true, 00:09:57.528 "unmap": true, 00:09:57.528 "flush": true, 00:09:57.528 "reset": true, 00:09:57.528 "nvme_admin": false, 00:09:57.528 "nvme_io": false, 00:09:57.528 "nvme_io_md": false, 00:09:57.528 "write_zeroes": true, 00:09:57.528 "zcopy": true, 00:09:57.528 "get_zone_info": false, 00:09:57.528 "zone_management": false, 00:09:57.528 "zone_append": false, 00:09:57.528 "compare": false, 00:09:57.528 "compare_and_write": false, 00:09:57.528 "abort": true, 00:09:57.528 "seek_hole": false, 00:09:57.528 "seek_data": false, 00:09:57.528 "copy": true, 00:09:57.528 "nvme_iov_md": false 00:09:57.528 }, 00:09:57.528 "memory_domains": [ 00:09:57.528 { 00:09:57.528 "dma_device_id": "system", 00:09:57.528 "dma_device_type": 1 00:09:57.528 }, 00:09:57.528 { 00:09:57.528 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:57.528 "dma_device_type": 2 00:09:57.528 } 00:09:57.528 ], 00:09:57.528 "driver_specific": {} 00:09:57.528 } 00:09:57.528 ] 00:09:57.528 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.528 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:57.528 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:57.528 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:57.528 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:57.528 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.528 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.787 BaseBdev3 00:09:57.787 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.787 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:09:57.787 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:09:57.787 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:57.787 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:57.787 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:57.787 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:57.787 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:57.787 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.787 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.787 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.787 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:57.787 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.787 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.787 [ 00:09:57.787 { 00:09:57.787 "name": "BaseBdev3", 00:09:57.787 "aliases": [ 00:09:57.787 "cb113325-c867-42cd-bba5-d0eb8666670d" 00:09:57.787 ], 00:09:57.787 "product_name": "Malloc disk", 00:09:57.787 "block_size": 512, 00:09:57.787 "num_blocks": 65536, 00:09:57.787 "uuid": "cb113325-c867-42cd-bba5-d0eb8666670d", 00:09:57.787 "assigned_rate_limits": { 00:09:57.787 "rw_ios_per_sec": 0, 00:09:57.787 "rw_mbytes_per_sec": 0, 00:09:57.787 "r_mbytes_per_sec": 0, 00:09:57.787 "w_mbytes_per_sec": 0 00:09:57.787 }, 00:09:57.787 "claimed": false, 00:09:57.787 "zoned": false, 00:09:57.787 "supported_io_types": { 00:09:57.787 "read": true, 00:09:57.787 "write": true, 00:09:57.787 "unmap": true, 00:09:57.787 "flush": true, 00:09:57.787 "reset": true, 00:09:57.787 "nvme_admin": false, 00:09:57.787 "nvme_io": false, 00:09:57.787 "nvme_io_md": false, 00:09:57.787 "write_zeroes": true, 00:09:57.787 "zcopy": true, 00:09:57.787 "get_zone_info": false, 00:09:57.787 "zone_management": false, 00:09:57.787 "zone_append": false, 00:09:57.787 "compare": false, 00:09:57.787 "compare_and_write": false, 00:09:57.787 "abort": true, 00:09:57.787 "seek_hole": false, 00:09:57.787 "seek_data": false, 00:09:57.787 "copy": true, 00:09:57.787 "nvme_iov_md": false 00:09:57.787 }, 00:09:57.787 "memory_domains": [ 00:09:57.787 { 00:09:57.787 "dma_device_id": "system", 00:09:57.787 "dma_device_type": 1 00:09:57.787 }, 00:09:57.787 { 00:09:57.787 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:57.787 "dma_device_type": 2 00:09:57.787 } 00:09:57.787 ], 00:09:57.787 "driver_specific": {} 00:09:57.787 } 00:09:57.787 ] 00:09:57.787 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.787 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:57.787 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:57.787 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:57.787 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:57.788 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.788 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.788 [2024-11-19 15:15:47.916186] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:57.788 [2024-11-19 15:15:47.916249] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:57.788 [2024-11-19 15:15:47.916276] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:57.788 [2024-11-19 15:15:47.918624] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:57.788 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.788 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:57.788 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:57.788 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:57.788 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:57.788 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:57.788 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:57.788 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:57.788 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:57.788 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:57.788 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:57.788 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:57.788 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:57.788 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.788 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:57.788 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.788 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:57.788 "name": "Existed_Raid", 00:09:57.788 "uuid": "fec14847-9afc-42d2-b7d0-e61690fd415d", 00:09:57.788 "strip_size_kb": 64, 00:09:57.788 "state": "configuring", 00:09:57.788 "raid_level": "concat", 00:09:57.788 "superblock": true, 00:09:57.788 "num_base_bdevs": 3, 00:09:57.788 "num_base_bdevs_discovered": 2, 00:09:57.788 "num_base_bdevs_operational": 3, 00:09:57.788 "base_bdevs_list": [ 00:09:57.788 { 00:09:57.788 "name": "BaseBdev1", 00:09:57.788 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:57.788 "is_configured": false, 00:09:57.788 "data_offset": 0, 00:09:57.788 "data_size": 0 00:09:57.788 }, 00:09:57.788 { 00:09:57.788 "name": "BaseBdev2", 00:09:57.788 "uuid": "6dff4074-657f-49cb-8046-a00c33ef1c1c", 00:09:57.788 "is_configured": true, 00:09:57.788 "data_offset": 2048, 00:09:57.788 "data_size": 63488 00:09:57.788 }, 00:09:57.788 { 00:09:57.788 "name": "BaseBdev3", 00:09:57.788 "uuid": "cb113325-c867-42cd-bba5-d0eb8666670d", 00:09:57.788 "is_configured": true, 00:09:57.788 "data_offset": 2048, 00:09:57.788 "data_size": 63488 00:09:57.788 } 00:09:57.788 ] 00:09:57.788 }' 00:09:57.788 15:15:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:57.788 15:15:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:58.046 15:15:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:09:58.046 15:15:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.046 15:15:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:58.046 [2024-11-19 15:15:48.339917] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:58.046 15:15:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.046 15:15:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:58.046 15:15:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:58.046 15:15:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:58.046 15:15:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:58.046 15:15:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:58.047 15:15:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:58.047 15:15:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:58.047 15:15:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:58.047 15:15:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:58.047 15:15:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:58.047 15:15:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:58.047 15:15:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:58.047 15:15:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.047 15:15:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:58.047 15:15:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.306 15:15:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:58.306 "name": "Existed_Raid", 00:09:58.306 "uuid": "fec14847-9afc-42d2-b7d0-e61690fd415d", 00:09:58.306 "strip_size_kb": 64, 00:09:58.306 "state": "configuring", 00:09:58.306 "raid_level": "concat", 00:09:58.306 "superblock": true, 00:09:58.306 "num_base_bdevs": 3, 00:09:58.306 "num_base_bdevs_discovered": 1, 00:09:58.306 "num_base_bdevs_operational": 3, 00:09:58.306 "base_bdevs_list": [ 00:09:58.306 { 00:09:58.306 "name": "BaseBdev1", 00:09:58.306 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:58.306 "is_configured": false, 00:09:58.306 "data_offset": 0, 00:09:58.306 "data_size": 0 00:09:58.306 }, 00:09:58.306 { 00:09:58.306 "name": null, 00:09:58.306 "uuid": "6dff4074-657f-49cb-8046-a00c33ef1c1c", 00:09:58.306 "is_configured": false, 00:09:58.306 "data_offset": 0, 00:09:58.306 "data_size": 63488 00:09:58.306 }, 00:09:58.306 { 00:09:58.306 "name": "BaseBdev3", 00:09:58.306 "uuid": "cb113325-c867-42cd-bba5-d0eb8666670d", 00:09:58.306 "is_configured": true, 00:09:58.306 "data_offset": 2048, 00:09:58.306 "data_size": 63488 00:09:58.306 } 00:09:58.306 ] 00:09:58.306 }' 00:09:58.306 15:15:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:58.306 15:15:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:58.589 15:15:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:58.589 15:15:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.589 15:15:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:58.589 15:15:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:58.589 15:15:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.589 15:15:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:09:58.589 15:15:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:58.589 15:15:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.589 15:15:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:58.589 [2024-11-19 15:15:48.864475] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:58.589 BaseBdev1 00:09:58.589 15:15:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.589 15:15:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:09:58.589 15:15:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:58.589 15:15:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:58.589 15:15:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:58.589 15:15:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:58.589 15:15:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:58.589 15:15:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:58.589 15:15:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.589 15:15:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:58.589 15:15:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.589 15:15:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:58.589 15:15:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.589 15:15:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:58.589 [ 00:09:58.589 { 00:09:58.589 "name": "BaseBdev1", 00:09:58.589 "aliases": [ 00:09:58.589 "44f29878-9234-4b92-88e0-e47b968a6b87" 00:09:58.589 ], 00:09:58.589 "product_name": "Malloc disk", 00:09:58.589 "block_size": 512, 00:09:58.589 "num_blocks": 65536, 00:09:58.589 "uuid": "44f29878-9234-4b92-88e0-e47b968a6b87", 00:09:58.589 "assigned_rate_limits": { 00:09:58.589 "rw_ios_per_sec": 0, 00:09:58.589 "rw_mbytes_per_sec": 0, 00:09:58.589 "r_mbytes_per_sec": 0, 00:09:58.589 "w_mbytes_per_sec": 0 00:09:58.589 }, 00:09:58.589 "claimed": true, 00:09:58.589 "claim_type": "exclusive_write", 00:09:58.589 "zoned": false, 00:09:58.589 "supported_io_types": { 00:09:58.589 "read": true, 00:09:58.589 "write": true, 00:09:58.589 "unmap": true, 00:09:58.589 "flush": true, 00:09:58.589 "reset": true, 00:09:58.589 "nvme_admin": false, 00:09:58.589 "nvme_io": false, 00:09:58.589 "nvme_io_md": false, 00:09:58.589 "write_zeroes": true, 00:09:58.589 "zcopy": true, 00:09:58.589 "get_zone_info": false, 00:09:58.589 "zone_management": false, 00:09:58.589 "zone_append": false, 00:09:58.589 "compare": false, 00:09:58.589 "compare_and_write": false, 00:09:58.589 "abort": true, 00:09:58.589 "seek_hole": false, 00:09:58.589 "seek_data": false, 00:09:58.589 "copy": true, 00:09:58.589 "nvme_iov_md": false 00:09:58.589 }, 00:09:58.589 "memory_domains": [ 00:09:58.589 { 00:09:58.589 "dma_device_id": "system", 00:09:58.589 "dma_device_type": 1 00:09:58.589 }, 00:09:58.589 { 00:09:58.589 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:58.589 "dma_device_type": 2 00:09:58.589 } 00:09:58.589 ], 00:09:58.589 "driver_specific": {} 00:09:58.589 } 00:09:58.589 ] 00:09:58.589 15:15:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.589 15:15:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:58.589 15:15:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:58.589 15:15:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:58.589 15:15:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:58.589 15:15:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:58.589 15:15:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:58.589 15:15:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:58.589 15:15:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:58.589 15:15:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:58.589 15:15:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:58.589 15:15:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:58.589 15:15:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:58.589 15:15:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:58.589 15:15:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.589 15:15:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:58.856 15:15:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.856 15:15:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:58.856 "name": "Existed_Raid", 00:09:58.856 "uuid": "fec14847-9afc-42d2-b7d0-e61690fd415d", 00:09:58.856 "strip_size_kb": 64, 00:09:58.856 "state": "configuring", 00:09:58.856 "raid_level": "concat", 00:09:58.856 "superblock": true, 00:09:58.856 "num_base_bdevs": 3, 00:09:58.856 "num_base_bdevs_discovered": 2, 00:09:58.856 "num_base_bdevs_operational": 3, 00:09:58.856 "base_bdevs_list": [ 00:09:58.856 { 00:09:58.856 "name": "BaseBdev1", 00:09:58.856 "uuid": "44f29878-9234-4b92-88e0-e47b968a6b87", 00:09:58.856 "is_configured": true, 00:09:58.856 "data_offset": 2048, 00:09:58.856 "data_size": 63488 00:09:58.856 }, 00:09:58.856 { 00:09:58.856 "name": null, 00:09:58.856 "uuid": "6dff4074-657f-49cb-8046-a00c33ef1c1c", 00:09:58.856 "is_configured": false, 00:09:58.856 "data_offset": 0, 00:09:58.856 "data_size": 63488 00:09:58.856 }, 00:09:58.856 { 00:09:58.856 "name": "BaseBdev3", 00:09:58.856 "uuid": "cb113325-c867-42cd-bba5-d0eb8666670d", 00:09:58.856 "is_configured": true, 00:09:58.856 "data_offset": 2048, 00:09:58.856 "data_size": 63488 00:09:58.856 } 00:09:58.856 ] 00:09:58.856 }' 00:09:58.856 15:15:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:58.856 15:15:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:59.115 15:15:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:59.115 15:15:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:59.115 15:15:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.115 15:15:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:59.115 15:15:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.115 15:15:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:09:59.115 15:15:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:09:59.115 15:15:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.115 15:15:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:59.115 [2024-11-19 15:15:49.431900] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:59.115 15:15:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.115 15:15:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:59.115 15:15:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:59.115 15:15:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:59.115 15:15:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:59.115 15:15:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:59.115 15:15:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:59.115 15:15:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:59.115 15:15:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:59.115 15:15:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:59.115 15:15:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:59.115 15:15:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:59.115 15:15:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.115 15:15:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:59.115 15:15:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:59.374 15:15:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.374 15:15:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:59.374 "name": "Existed_Raid", 00:09:59.374 "uuid": "fec14847-9afc-42d2-b7d0-e61690fd415d", 00:09:59.374 "strip_size_kb": 64, 00:09:59.374 "state": "configuring", 00:09:59.374 "raid_level": "concat", 00:09:59.374 "superblock": true, 00:09:59.374 "num_base_bdevs": 3, 00:09:59.374 "num_base_bdevs_discovered": 1, 00:09:59.374 "num_base_bdevs_operational": 3, 00:09:59.374 "base_bdevs_list": [ 00:09:59.374 { 00:09:59.374 "name": "BaseBdev1", 00:09:59.374 "uuid": "44f29878-9234-4b92-88e0-e47b968a6b87", 00:09:59.374 "is_configured": true, 00:09:59.374 "data_offset": 2048, 00:09:59.374 "data_size": 63488 00:09:59.374 }, 00:09:59.374 { 00:09:59.374 "name": null, 00:09:59.374 "uuid": "6dff4074-657f-49cb-8046-a00c33ef1c1c", 00:09:59.374 "is_configured": false, 00:09:59.374 "data_offset": 0, 00:09:59.374 "data_size": 63488 00:09:59.374 }, 00:09:59.374 { 00:09:59.374 "name": null, 00:09:59.374 "uuid": "cb113325-c867-42cd-bba5-d0eb8666670d", 00:09:59.374 "is_configured": false, 00:09:59.374 "data_offset": 0, 00:09:59.375 "data_size": 63488 00:09:59.375 } 00:09:59.375 ] 00:09:59.375 }' 00:09:59.375 15:15:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:59.375 15:15:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:59.634 15:15:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:59.634 15:15:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:59.634 15:15:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.634 15:15:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:59.634 15:15:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.634 15:15:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:09:59.634 15:15:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:09:59.634 15:15:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.634 15:15:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:59.634 [2024-11-19 15:15:49.875324] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:59.634 15:15:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.634 15:15:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:59.634 15:15:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:59.634 15:15:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:59.634 15:15:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:59.634 15:15:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:59.634 15:15:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:59.634 15:15:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:59.634 15:15:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:59.634 15:15:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:59.634 15:15:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:59.634 15:15:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:59.634 15:15:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:59.634 15:15:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:59.634 15:15:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:59.634 15:15:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:59.634 15:15:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:59.634 "name": "Existed_Raid", 00:09:59.634 "uuid": "fec14847-9afc-42d2-b7d0-e61690fd415d", 00:09:59.634 "strip_size_kb": 64, 00:09:59.634 "state": "configuring", 00:09:59.634 "raid_level": "concat", 00:09:59.634 "superblock": true, 00:09:59.634 "num_base_bdevs": 3, 00:09:59.634 "num_base_bdevs_discovered": 2, 00:09:59.634 "num_base_bdevs_operational": 3, 00:09:59.634 "base_bdevs_list": [ 00:09:59.634 { 00:09:59.634 "name": "BaseBdev1", 00:09:59.634 "uuid": "44f29878-9234-4b92-88e0-e47b968a6b87", 00:09:59.634 "is_configured": true, 00:09:59.634 "data_offset": 2048, 00:09:59.634 "data_size": 63488 00:09:59.634 }, 00:09:59.634 { 00:09:59.634 "name": null, 00:09:59.634 "uuid": "6dff4074-657f-49cb-8046-a00c33ef1c1c", 00:09:59.634 "is_configured": false, 00:09:59.634 "data_offset": 0, 00:09:59.634 "data_size": 63488 00:09:59.634 }, 00:09:59.634 { 00:09:59.634 "name": "BaseBdev3", 00:09:59.634 "uuid": "cb113325-c867-42cd-bba5-d0eb8666670d", 00:09:59.634 "is_configured": true, 00:09:59.634 "data_offset": 2048, 00:09:59.634 "data_size": 63488 00:09:59.634 } 00:09:59.634 ] 00:09:59.634 }' 00:09:59.634 15:15:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:59.634 15:15:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.203 15:15:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:00.203 15:15:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:00.203 15:15:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.203 15:15:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.203 15:15:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.203 15:15:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:10:00.203 15:15:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:00.203 15:15:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.203 15:15:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.203 [2024-11-19 15:15:50.382536] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:00.203 15:15:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.203 15:15:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:10:00.203 15:15:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:00.203 15:15:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:00.203 15:15:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:00.203 15:15:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:00.203 15:15:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:00.203 15:15:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:00.203 15:15:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:00.203 15:15:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:00.203 15:15:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:00.203 15:15:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:00.203 15:15:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:00.203 15:15:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.203 15:15:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.203 15:15:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.203 15:15:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:00.203 "name": "Existed_Raid", 00:10:00.203 "uuid": "fec14847-9afc-42d2-b7d0-e61690fd415d", 00:10:00.203 "strip_size_kb": 64, 00:10:00.203 "state": "configuring", 00:10:00.203 "raid_level": "concat", 00:10:00.203 "superblock": true, 00:10:00.203 "num_base_bdevs": 3, 00:10:00.203 "num_base_bdevs_discovered": 1, 00:10:00.203 "num_base_bdevs_operational": 3, 00:10:00.203 "base_bdevs_list": [ 00:10:00.203 { 00:10:00.203 "name": null, 00:10:00.203 "uuid": "44f29878-9234-4b92-88e0-e47b968a6b87", 00:10:00.203 "is_configured": false, 00:10:00.203 "data_offset": 0, 00:10:00.203 "data_size": 63488 00:10:00.203 }, 00:10:00.203 { 00:10:00.203 "name": null, 00:10:00.203 "uuid": "6dff4074-657f-49cb-8046-a00c33ef1c1c", 00:10:00.203 "is_configured": false, 00:10:00.203 "data_offset": 0, 00:10:00.203 "data_size": 63488 00:10:00.203 }, 00:10:00.203 { 00:10:00.203 "name": "BaseBdev3", 00:10:00.203 "uuid": "cb113325-c867-42cd-bba5-d0eb8666670d", 00:10:00.203 "is_configured": true, 00:10:00.203 "data_offset": 2048, 00:10:00.203 "data_size": 63488 00:10:00.203 } 00:10:00.203 ] 00:10:00.203 }' 00:10:00.203 15:15:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:00.203 15:15:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.772 15:15:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:00.772 15:15:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.772 15:15:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.772 15:15:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:00.772 15:15:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.772 15:15:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:10:00.772 15:15:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:10:00.772 15:15:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.772 15:15:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.772 [2024-11-19 15:15:50.913873] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:00.772 15:15:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.772 15:15:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:10:00.772 15:15:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:00.772 15:15:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:00.772 15:15:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:00.772 15:15:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:00.772 15:15:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:00.772 15:15:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:00.772 15:15:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:00.772 15:15:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:00.772 15:15:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:00.772 15:15:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:00.772 15:15:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:00.772 15:15:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.772 15:15:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.772 15:15:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.772 15:15:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:00.772 "name": "Existed_Raid", 00:10:00.772 "uuid": "fec14847-9afc-42d2-b7d0-e61690fd415d", 00:10:00.772 "strip_size_kb": 64, 00:10:00.772 "state": "configuring", 00:10:00.772 "raid_level": "concat", 00:10:00.772 "superblock": true, 00:10:00.772 "num_base_bdevs": 3, 00:10:00.772 "num_base_bdevs_discovered": 2, 00:10:00.772 "num_base_bdevs_operational": 3, 00:10:00.772 "base_bdevs_list": [ 00:10:00.772 { 00:10:00.772 "name": null, 00:10:00.772 "uuid": "44f29878-9234-4b92-88e0-e47b968a6b87", 00:10:00.772 "is_configured": false, 00:10:00.772 "data_offset": 0, 00:10:00.772 "data_size": 63488 00:10:00.772 }, 00:10:00.772 { 00:10:00.772 "name": "BaseBdev2", 00:10:00.772 "uuid": "6dff4074-657f-49cb-8046-a00c33ef1c1c", 00:10:00.772 "is_configured": true, 00:10:00.772 "data_offset": 2048, 00:10:00.772 "data_size": 63488 00:10:00.772 }, 00:10:00.772 { 00:10:00.772 "name": "BaseBdev3", 00:10:00.772 "uuid": "cb113325-c867-42cd-bba5-d0eb8666670d", 00:10:00.772 "is_configured": true, 00:10:00.772 "data_offset": 2048, 00:10:00.772 "data_size": 63488 00:10:00.772 } 00:10:00.772 ] 00:10:00.772 }' 00:10:00.772 15:15:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:00.772 15:15:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.032 15:15:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:01.032 15:15:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.032 15:15:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.032 15:15:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:01.032 15:15:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.032 15:15:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:10:01.032 15:15:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:10:01.032 15:15:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:01.032 15:15:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.032 15:15:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.292 15:15:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.292 15:15:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 44f29878-9234-4b92-88e0-e47b968a6b87 00:10:01.292 15:15:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.292 15:15:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.292 [2024-11-19 15:15:51.413930] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:10:01.292 [2024-11-19 15:15:51.414149] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:10:01.292 [2024-11-19 15:15:51.414168] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:10:01.292 [2024-11-19 15:15:51.414450] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:10:01.293 [2024-11-19 15:15:51.414586] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:10:01.293 [2024-11-19 15:15:51.414603] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:10:01.293 NewBaseBdev 00:10:01.293 [2024-11-19 15:15:51.414725] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:01.293 15:15:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.293 15:15:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:10:01.293 15:15:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:10:01.293 15:15:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:01.293 15:15:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:01.293 15:15:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:01.293 15:15:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:01.293 15:15:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:01.293 15:15:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.293 15:15:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.293 15:15:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.293 15:15:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:10:01.293 15:15:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.293 15:15:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.293 [ 00:10:01.293 { 00:10:01.293 "name": "NewBaseBdev", 00:10:01.293 "aliases": [ 00:10:01.293 "44f29878-9234-4b92-88e0-e47b968a6b87" 00:10:01.293 ], 00:10:01.293 "product_name": "Malloc disk", 00:10:01.293 "block_size": 512, 00:10:01.293 "num_blocks": 65536, 00:10:01.293 "uuid": "44f29878-9234-4b92-88e0-e47b968a6b87", 00:10:01.293 "assigned_rate_limits": { 00:10:01.293 "rw_ios_per_sec": 0, 00:10:01.293 "rw_mbytes_per_sec": 0, 00:10:01.293 "r_mbytes_per_sec": 0, 00:10:01.293 "w_mbytes_per_sec": 0 00:10:01.293 }, 00:10:01.293 "claimed": true, 00:10:01.293 "claim_type": "exclusive_write", 00:10:01.293 "zoned": false, 00:10:01.293 "supported_io_types": { 00:10:01.293 "read": true, 00:10:01.293 "write": true, 00:10:01.293 "unmap": true, 00:10:01.293 "flush": true, 00:10:01.293 "reset": true, 00:10:01.293 "nvme_admin": false, 00:10:01.293 "nvme_io": false, 00:10:01.293 "nvme_io_md": false, 00:10:01.293 "write_zeroes": true, 00:10:01.293 "zcopy": true, 00:10:01.293 "get_zone_info": false, 00:10:01.293 "zone_management": false, 00:10:01.293 "zone_append": false, 00:10:01.293 "compare": false, 00:10:01.293 "compare_and_write": false, 00:10:01.293 "abort": true, 00:10:01.293 "seek_hole": false, 00:10:01.293 "seek_data": false, 00:10:01.293 "copy": true, 00:10:01.293 "nvme_iov_md": false 00:10:01.293 }, 00:10:01.293 "memory_domains": [ 00:10:01.293 { 00:10:01.293 "dma_device_id": "system", 00:10:01.293 "dma_device_type": 1 00:10:01.293 }, 00:10:01.293 { 00:10:01.293 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:01.293 "dma_device_type": 2 00:10:01.293 } 00:10:01.293 ], 00:10:01.293 "driver_specific": {} 00:10:01.293 } 00:10:01.293 ] 00:10:01.293 15:15:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.293 15:15:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:01.293 15:15:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:10:01.293 15:15:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:01.293 15:15:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:01.293 15:15:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:01.293 15:15:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:01.293 15:15:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:01.293 15:15:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:01.293 15:15:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:01.293 15:15:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:01.293 15:15:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:01.293 15:15:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:01.293 15:15:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.293 15:15:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.293 15:15:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:01.293 15:15:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.293 15:15:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:01.293 "name": "Existed_Raid", 00:10:01.293 "uuid": "fec14847-9afc-42d2-b7d0-e61690fd415d", 00:10:01.293 "strip_size_kb": 64, 00:10:01.293 "state": "online", 00:10:01.293 "raid_level": "concat", 00:10:01.293 "superblock": true, 00:10:01.293 "num_base_bdevs": 3, 00:10:01.293 "num_base_bdevs_discovered": 3, 00:10:01.293 "num_base_bdevs_operational": 3, 00:10:01.293 "base_bdevs_list": [ 00:10:01.293 { 00:10:01.293 "name": "NewBaseBdev", 00:10:01.293 "uuid": "44f29878-9234-4b92-88e0-e47b968a6b87", 00:10:01.293 "is_configured": true, 00:10:01.293 "data_offset": 2048, 00:10:01.293 "data_size": 63488 00:10:01.293 }, 00:10:01.293 { 00:10:01.293 "name": "BaseBdev2", 00:10:01.293 "uuid": "6dff4074-657f-49cb-8046-a00c33ef1c1c", 00:10:01.293 "is_configured": true, 00:10:01.293 "data_offset": 2048, 00:10:01.293 "data_size": 63488 00:10:01.293 }, 00:10:01.293 { 00:10:01.293 "name": "BaseBdev3", 00:10:01.293 "uuid": "cb113325-c867-42cd-bba5-d0eb8666670d", 00:10:01.293 "is_configured": true, 00:10:01.293 "data_offset": 2048, 00:10:01.293 "data_size": 63488 00:10:01.293 } 00:10:01.293 ] 00:10:01.293 }' 00:10:01.293 15:15:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:01.293 15:15:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.552 15:15:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:10:01.552 15:15:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:01.552 15:15:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:01.552 15:15:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:01.553 15:15:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:10:01.553 15:15:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:01.553 15:15:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:01.553 15:15:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:01.553 15:15:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.553 15:15:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.553 [2024-11-19 15:15:51.837687] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:01.553 15:15:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.553 15:15:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:01.553 "name": "Existed_Raid", 00:10:01.553 "aliases": [ 00:10:01.553 "fec14847-9afc-42d2-b7d0-e61690fd415d" 00:10:01.553 ], 00:10:01.553 "product_name": "Raid Volume", 00:10:01.553 "block_size": 512, 00:10:01.553 "num_blocks": 190464, 00:10:01.553 "uuid": "fec14847-9afc-42d2-b7d0-e61690fd415d", 00:10:01.553 "assigned_rate_limits": { 00:10:01.553 "rw_ios_per_sec": 0, 00:10:01.553 "rw_mbytes_per_sec": 0, 00:10:01.553 "r_mbytes_per_sec": 0, 00:10:01.553 "w_mbytes_per_sec": 0 00:10:01.553 }, 00:10:01.553 "claimed": false, 00:10:01.553 "zoned": false, 00:10:01.553 "supported_io_types": { 00:10:01.553 "read": true, 00:10:01.553 "write": true, 00:10:01.553 "unmap": true, 00:10:01.553 "flush": true, 00:10:01.553 "reset": true, 00:10:01.553 "nvme_admin": false, 00:10:01.553 "nvme_io": false, 00:10:01.553 "nvme_io_md": false, 00:10:01.553 "write_zeroes": true, 00:10:01.553 "zcopy": false, 00:10:01.553 "get_zone_info": false, 00:10:01.553 "zone_management": false, 00:10:01.553 "zone_append": false, 00:10:01.553 "compare": false, 00:10:01.553 "compare_and_write": false, 00:10:01.553 "abort": false, 00:10:01.553 "seek_hole": false, 00:10:01.553 "seek_data": false, 00:10:01.553 "copy": false, 00:10:01.553 "nvme_iov_md": false 00:10:01.553 }, 00:10:01.553 "memory_domains": [ 00:10:01.553 { 00:10:01.553 "dma_device_id": "system", 00:10:01.553 "dma_device_type": 1 00:10:01.553 }, 00:10:01.553 { 00:10:01.553 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:01.553 "dma_device_type": 2 00:10:01.553 }, 00:10:01.553 { 00:10:01.553 "dma_device_id": "system", 00:10:01.553 "dma_device_type": 1 00:10:01.553 }, 00:10:01.553 { 00:10:01.553 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:01.553 "dma_device_type": 2 00:10:01.553 }, 00:10:01.553 { 00:10:01.553 "dma_device_id": "system", 00:10:01.553 "dma_device_type": 1 00:10:01.553 }, 00:10:01.553 { 00:10:01.553 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:01.553 "dma_device_type": 2 00:10:01.553 } 00:10:01.553 ], 00:10:01.553 "driver_specific": { 00:10:01.553 "raid": { 00:10:01.553 "uuid": "fec14847-9afc-42d2-b7d0-e61690fd415d", 00:10:01.553 "strip_size_kb": 64, 00:10:01.553 "state": "online", 00:10:01.553 "raid_level": "concat", 00:10:01.553 "superblock": true, 00:10:01.553 "num_base_bdevs": 3, 00:10:01.553 "num_base_bdevs_discovered": 3, 00:10:01.553 "num_base_bdevs_operational": 3, 00:10:01.553 "base_bdevs_list": [ 00:10:01.553 { 00:10:01.553 "name": "NewBaseBdev", 00:10:01.553 "uuid": "44f29878-9234-4b92-88e0-e47b968a6b87", 00:10:01.553 "is_configured": true, 00:10:01.553 "data_offset": 2048, 00:10:01.553 "data_size": 63488 00:10:01.553 }, 00:10:01.553 { 00:10:01.553 "name": "BaseBdev2", 00:10:01.553 "uuid": "6dff4074-657f-49cb-8046-a00c33ef1c1c", 00:10:01.553 "is_configured": true, 00:10:01.553 "data_offset": 2048, 00:10:01.553 "data_size": 63488 00:10:01.553 }, 00:10:01.553 { 00:10:01.553 "name": "BaseBdev3", 00:10:01.553 "uuid": "cb113325-c867-42cd-bba5-d0eb8666670d", 00:10:01.553 "is_configured": true, 00:10:01.553 "data_offset": 2048, 00:10:01.553 "data_size": 63488 00:10:01.553 } 00:10:01.553 ] 00:10:01.553 } 00:10:01.553 } 00:10:01.553 }' 00:10:01.553 15:15:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:01.813 15:15:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:10:01.813 BaseBdev2 00:10:01.813 BaseBdev3' 00:10:01.813 15:15:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:01.813 15:15:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:01.813 15:15:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:01.813 15:15:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:01.813 15:15:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:10:01.813 15:15:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.813 15:15:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.813 15:15:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.813 15:15:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:01.813 15:15:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:01.813 15:15:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:01.813 15:15:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:01.813 15:15:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.813 15:15:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.813 15:15:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:01.813 15:15:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.813 15:15:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:01.813 15:15:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:01.813 15:15:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:01.813 15:15:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:01.813 15:15:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:01.813 15:15:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.813 15:15:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.813 15:15:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.813 15:15:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:01.813 15:15:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:01.813 15:15:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:01.813 15:15:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.813 15:15:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.813 [2024-11-19 15:15:52.088963] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:01.813 [2024-11-19 15:15:52.089022] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:01.813 [2024-11-19 15:15:52.089131] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:01.813 [2024-11-19 15:15:52.089198] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:01.813 [2024-11-19 15:15:52.089213] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:10:01.813 15:15:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.813 15:15:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 77341 00:10:01.813 15:15:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 77341 ']' 00:10:01.813 15:15:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 77341 00:10:01.813 15:15:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:10:01.813 15:15:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:01.813 15:15:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 77341 00:10:01.813 15:15:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:01.813 15:15:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:01.813 killing process with pid 77341 00:10:01.813 15:15:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 77341' 00:10:01.813 15:15:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 77341 00:10:01.813 [2024-11-19 15:15:52.137736] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:01.813 15:15:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 77341 00:10:02.073 [2024-11-19 15:15:52.199310] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:02.332 15:15:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:10:02.333 00:10:02.333 real 0m9.063s 00:10:02.333 user 0m15.221s 00:10:02.333 sys 0m1.864s 00:10:02.333 15:15:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:02.333 15:15:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:02.333 ************************************ 00:10:02.333 END TEST raid_state_function_test_sb 00:10:02.333 ************************************ 00:10:02.333 15:15:52 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test concat 3 00:10:02.333 15:15:52 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:10:02.333 15:15:52 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:02.333 15:15:52 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:02.333 ************************************ 00:10:02.333 START TEST raid_superblock_test 00:10:02.333 ************************************ 00:10:02.333 15:15:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test concat 3 00:10:02.333 15:15:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=concat 00:10:02.333 15:15:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:10:02.333 15:15:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:10:02.333 15:15:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:10:02.333 15:15:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:10:02.333 15:15:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:10:02.333 15:15:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:10:02.333 15:15:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:10:02.333 15:15:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:10:02.333 15:15:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:10:02.333 15:15:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:10:02.333 15:15:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:10:02.333 15:15:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:10:02.333 15:15:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' concat '!=' raid1 ']' 00:10:02.333 15:15:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:10:02.333 15:15:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:10:02.333 15:15:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=77950 00:10:02.333 15:15:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 77950 00:10:02.333 15:15:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:10:02.333 15:15:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 77950 ']' 00:10:02.333 15:15:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:02.333 15:15:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:02.333 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:02.333 15:15:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:02.333 15:15:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:02.333 15:15:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.592 [2024-11-19 15:15:52.700816] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:10:02.592 [2024-11-19 15:15:52.700982] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid77950 ] 00:10:02.592 [2024-11-19 15:15:52.859805] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:02.592 [2024-11-19 15:15:52.900958] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:02.852 [2024-11-19 15:15:52.978065] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:02.852 [2024-11-19 15:15:52.978113] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:03.421 15:15:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:03.421 15:15:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:10:03.421 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:10:03.421 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:03.421 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:10:03.421 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:10:03.421 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:10:03.421 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:03.421 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:03.421 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:03.421 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:10:03.421 15:15:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.421 15:15:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.421 malloc1 00:10:03.421 15:15:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.421 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:03.421 15:15:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.421 15:15:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.421 [2024-11-19 15:15:53.568997] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:03.421 [2024-11-19 15:15:53.569069] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:03.421 [2024-11-19 15:15:53.569088] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:10:03.421 [2024-11-19 15:15:53.569112] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:03.421 [2024-11-19 15:15:53.571652] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:03.421 [2024-11-19 15:15:53.571692] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:03.421 pt1 00:10:03.421 15:15:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.421 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:03.421 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:03.421 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:10:03.421 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:10:03.421 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:10:03.421 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:03.421 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:03.421 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:03.421 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:10:03.421 15:15:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.421 15:15:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.421 malloc2 00:10:03.421 15:15:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.421 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:03.421 15:15:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.421 15:15:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.421 [2024-11-19 15:15:53.608208] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:03.421 [2024-11-19 15:15:53.608274] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:03.421 [2024-11-19 15:15:53.608292] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:10:03.421 [2024-11-19 15:15:53.608306] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:03.421 [2024-11-19 15:15:53.611071] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:03.421 [2024-11-19 15:15:53.611110] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:03.421 pt2 00:10:03.421 15:15:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.421 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:03.421 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:03.422 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:10:03.422 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:10:03.422 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:10:03.422 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:03.422 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:03.422 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:03.422 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:10:03.422 15:15:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.422 15:15:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.422 malloc3 00:10:03.422 15:15:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.422 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:03.422 15:15:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.422 15:15:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.422 [2024-11-19 15:15:53.643315] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:03.422 [2024-11-19 15:15:53.643387] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:03.422 [2024-11-19 15:15:53.643408] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:10:03.422 [2024-11-19 15:15:53.643421] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:03.422 [2024-11-19 15:15:53.646205] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:03.422 [2024-11-19 15:15:53.646244] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:03.422 pt3 00:10:03.422 15:15:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.422 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:03.422 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:03.422 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:10:03.422 15:15:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.422 15:15:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.422 [2024-11-19 15:15:53.655366] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:03.422 [2024-11-19 15:15:53.657855] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:03.422 [2024-11-19 15:15:53.657922] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:03.422 [2024-11-19 15:15:53.658113] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:10:03.422 [2024-11-19 15:15:53.658128] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:10:03.422 [2024-11-19 15:15:53.658458] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:10:03.422 [2024-11-19 15:15:53.658629] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:10:03.422 [2024-11-19 15:15:53.658659] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:10:03.422 [2024-11-19 15:15:53.658820] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:03.422 15:15:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.422 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:10:03.422 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:03.422 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:03.422 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:03.422 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:03.422 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:03.422 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:03.422 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:03.422 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:03.422 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:03.422 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:03.422 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:03.422 15:15:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.422 15:15:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.422 15:15:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.422 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:03.422 "name": "raid_bdev1", 00:10:03.422 "uuid": "ea435670-ddbf-4a54-8fe3-340b4dab47c2", 00:10:03.422 "strip_size_kb": 64, 00:10:03.422 "state": "online", 00:10:03.422 "raid_level": "concat", 00:10:03.422 "superblock": true, 00:10:03.422 "num_base_bdevs": 3, 00:10:03.422 "num_base_bdevs_discovered": 3, 00:10:03.422 "num_base_bdevs_operational": 3, 00:10:03.422 "base_bdevs_list": [ 00:10:03.422 { 00:10:03.422 "name": "pt1", 00:10:03.422 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:03.422 "is_configured": true, 00:10:03.422 "data_offset": 2048, 00:10:03.422 "data_size": 63488 00:10:03.422 }, 00:10:03.422 { 00:10:03.422 "name": "pt2", 00:10:03.422 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:03.422 "is_configured": true, 00:10:03.422 "data_offset": 2048, 00:10:03.422 "data_size": 63488 00:10:03.422 }, 00:10:03.422 { 00:10:03.422 "name": "pt3", 00:10:03.422 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:03.422 "is_configured": true, 00:10:03.422 "data_offset": 2048, 00:10:03.422 "data_size": 63488 00:10:03.422 } 00:10:03.422 ] 00:10:03.422 }' 00:10:03.422 15:15:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:03.422 15:15:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.682 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:10:03.682 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:10:03.682 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:03.682 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:03.682 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:03.682 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:03.682 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:03.682 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:03.682 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.682 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.682 [2024-11-19 15:15:54.015133] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:03.942 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.942 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:03.942 "name": "raid_bdev1", 00:10:03.942 "aliases": [ 00:10:03.942 "ea435670-ddbf-4a54-8fe3-340b4dab47c2" 00:10:03.942 ], 00:10:03.942 "product_name": "Raid Volume", 00:10:03.942 "block_size": 512, 00:10:03.942 "num_blocks": 190464, 00:10:03.942 "uuid": "ea435670-ddbf-4a54-8fe3-340b4dab47c2", 00:10:03.942 "assigned_rate_limits": { 00:10:03.942 "rw_ios_per_sec": 0, 00:10:03.942 "rw_mbytes_per_sec": 0, 00:10:03.942 "r_mbytes_per_sec": 0, 00:10:03.942 "w_mbytes_per_sec": 0 00:10:03.942 }, 00:10:03.942 "claimed": false, 00:10:03.942 "zoned": false, 00:10:03.942 "supported_io_types": { 00:10:03.942 "read": true, 00:10:03.942 "write": true, 00:10:03.942 "unmap": true, 00:10:03.942 "flush": true, 00:10:03.942 "reset": true, 00:10:03.942 "nvme_admin": false, 00:10:03.942 "nvme_io": false, 00:10:03.942 "nvme_io_md": false, 00:10:03.942 "write_zeroes": true, 00:10:03.942 "zcopy": false, 00:10:03.942 "get_zone_info": false, 00:10:03.942 "zone_management": false, 00:10:03.942 "zone_append": false, 00:10:03.942 "compare": false, 00:10:03.942 "compare_and_write": false, 00:10:03.942 "abort": false, 00:10:03.942 "seek_hole": false, 00:10:03.942 "seek_data": false, 00:10:03.942 "copy": false, 00:10:03.942 "nvme_iov_md": false 00:10:03.942 }, 00:10:03.942 "memory_domains": [ 00:10:03.942 { 00:10:03.942 "dma_device_id": "system", 00:10:03.942 "dma_device_type": 1 00:10:03.942 }, 00:10:03.942 { 00:10:03.942 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:03.942 "dma_device_type": 2 00:10:03.942 }, 00:10:03.942 { 00:10:03.942 "dma_device_id": "system", 00:10:03.942 "dma_device_type": 1 00:10:03.942 }, 00:10:03.942 { 00:10:03.942 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:03.942 "dma_device_type": 2 00:10:03.942 }, 00:10:03.942 { 00:10:03.942 "dma_device_id": "system", 00:10:03.942 "dma_device_type": 1 00:10:03.942 }, 00:10:03.942 { 00:10:03.942 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:03.942 "dma_device_type": 2 00:10:03.942 } 00:10:03.942 ], 00:10:03.942 "driver_specific": { 00:10:03.942 "raid": { 00:10:03.942 "uuid": "ea435670-ddbf-4a54-8fe3-340b4dab47c2", 00:10:03.942 "strip_size_kb": 64, 00:10:03.942 "state": "online", 00:10:03.942 "raid_level": "concat", 00:10:03.942 "superblock": true, 00:10:03.942 "num_base_bdevs": 3, 00:10:03.942 "num_base_bdevs_discovered": 3, 00:10:03.942 "num_base_bdevs_operational": 3, 00:10:03.942 "base_bdevs_list": [ 00:10:03.942 { 00:10:03.942 "name": "pt1", 00:10:03.942 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:03.942 "is_configured": true, 00:10:03.942 "data_offset": 2048, 00:10:03.942 "data_size": 63488 00:10:03.942 }, 00:10:03.942 { 00:10:03.942 "name": "pt2", 00:10:03.942 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:03.942 "is_configured": true, 00:10:03.942 "data_offset": 2048, 00:10:03.942 "data_size": 63488 00:10:03.942 }, 00:10:03.942 { 00:10:03.942 "name": "pt3", 00:10:03.942 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:03.942 "is_configured": true, 00:10:03.942 "data_offset": 2048, 00:10:03.942 "data_size": 63488 00:10:03.942 } 00:10:03.942 ] 00:10:03.942 } 00:10:03.942 } 00:10:03.942 }' 00:10:03.942 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:03.942 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:10:03.942 pt2 00:10:03.942 pt3' 00:10:03.942 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:03.942 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:03.942 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:03.943 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:03.943 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:10:03.943 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.943 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.943 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.943 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:03.943 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:03.943 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:03.943 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:10:03.943 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:03.943 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.943 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.943 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.943 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:03.943 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:03.943 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:03.943 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:03.943 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:10:03.943 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.943 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.943 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.203 [2024-11-19 15:15:54.294532] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=ea435670-ddbf-4a54-8fe3-340b4dab47c2 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z ea435670-ddbf-4a54-8fe3-340b4dab47c2 ']' 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.203 [2024-11-19 15:15:54.326182] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:04.203 [2024-11-19 15:15:54.326228] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:04.203 [2024-11-19 15:15:54.326332] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:04.203 [2024-11-19 15:15:54.326408] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:04.203 [2024-11-19 15:15:54.326427] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:10:04.203 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:10:04.204 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:10:04.204 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:10:04.204 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.204 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.204 [2024-11-19 15:15:54.470035] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:10:04.204 [2024-11-19 15:15:54.472384] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:10:04.204 [2024-11-19 15:15:54.472437] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:10:04.204 [2024-11-19 15:15:54.472494] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:10:04.204 [2024-11-19 15:15:54.472546] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:10:04.204 [2024-11-19 15:15:54.472581] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:10:04.204 [2024-11-19 15:15:54.472600] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:04.204 [2024-11-19 15:15:54.472612] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:10:04.204 request: 00:10:04.204 { 00:10:04.204 "name": "raid_bdev1", 00:10:04.204 "raid_level": "concat", 00:10:04.204 "base_bdevs": [ 00:10:04.204 "malloc1", 00:10:04.204 "malloc2", 00:10:04.204 "malloc3" 00:10:04.204 ], 00:10:04.204 "strip_size_kb": 64, 00:10:04.204 "superblock": false, 00:10:04.204 "method": "bdev_raid_create", 00:10:04.204 "req_id": 1 00:10:04.204 } 00:10:04.204 Got JSON-RPC error response 00:10:04.204 response: 00:10:04.204 { 00:10:04.204 "code": -17, 00:10:04.204 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:10:04.204 } 00:10:04.204 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:10:04.204 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:10:04.204 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:10:04.204 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:10:04.204 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:10:04.204 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:04.204 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.204 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.204 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:10:04.204 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.204 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:10:04.204 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:10:04.204 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:04.204 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.204 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.204 [2024-11-19 15:15:54.537864] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:04.204 [2024-11-19 15:15:54.537942] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:04.204 [2024-11-19 15:15:54.537977] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:10:04.204 [2024-11-19 15:15:54.537992] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:04.464 [2024-11-19 15:15:54.541000] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:04.464 [2024-11-19 15:15:54.541041] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:04.464 [2024-11-19 15:15:54.541127] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:10:04.464 [2024-11-19 15:15:54.541170] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:04.464 pt1 00:10:04.464 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.464 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 3 00:10:04.464 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:04.464 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:04.464 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:04.464 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:04.464 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:04.464 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:04.464 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:04.464 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:04.464 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:04.464 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:04.464 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:04.464 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.464 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.464 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.464 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:04.464 "name": "raid_bdev1", 00:10:04.464 "uuid": "ea435670-ddbf-4a54-8fe3-340b4dab47c2", 00:10:04.464 "strip_size_kb": 64, 00:10:04.464 "state": "configuring", 00:10:04.464 "raid_level": "concat", 00:10:04.464 "superblock": true, 00:10:04.464 "num_base_bdevs": 3, 00:10:04.464 "num_base_bdevs_discovered": 1, 00:10:04.464 "num_base_bdevs_operational": 3, 00:10:04.464 "base_bdevs_list": [ 00:10:04.464 { 00:10:04.464 "name": "pt1", 00:10:04.464 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:04.464 "is_configured": true, 00:10:04.464 "data_offset": 2048, 00:10:04.464 "data_size": 63488 00:10:04.464 }, 00:10:04.464 { 00:10:04.464 "name": null, 00:10:04.464 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:04.464 "is_configured": false, 00:10:04.464 "data_offset": 2048, 00:10:04.464 "data_size": 63488 00:10:04.464 }, 00:10:04.464 { 00:10:04.464 "name": null, 00:10:04.464 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:04.464 "is_configured": false, 00:10:04.464 "data_offset": 2048, 00:10:04.464 "data_size": 63488 00:10:04.464 } 00:10:04.464 ] 00:10:04.464 }' 00:10:04.464 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:04.464 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.723 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:10:04.723 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:04.723 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.723 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.723 [2024-11-19 15:15:54.993222] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:04.723 [2024-11-19 15:15:54.993347] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:04.723 [2024-11-19 15:15:54.993376] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:10:04.723 [2024-11-19 15:15:54.993393] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:04.724 [2024-11-19 15:15:54.993932] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:04.724 [2024-11-19 15:15:54.993978] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:04.724 [2024-11-19 15:15:54.994078] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:04.724 [2024-11-19 15:15:54.994116] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:04.724 pt2 00:10:04.724 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.724 15:15:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:10:04.724 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.724 15:15:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.724 [2024-11-19 15:15:55.001168] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:10:04.724 15:15:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.724 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 3 00:10:04.724 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:04.724 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:04.724 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:04.724 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:04.724 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:04.724 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:04.724 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:04.724 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:04.724 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:04.724 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:04.724 15:15:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.724 15:15:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.724 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:04.724 15:15:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.724 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:04.724 "name": "raid_bdev1", 00:10:04.724 "uuid": "ea435670-ddbf-4a54-8fe3-340b4dab47c2", 00:10:04.724 "strip_size_kb": 64, 00:10:04.724 "state": "configuring", 00:10:04.724 "raid_level": "concat", 00:10:04.724 "superblock": true, 00:10:04.724 "num_base_bdevs": 3, 00:10:04.724 "num_base_bdevs_discovered": 1, 00:10:04.724 "num_base_bdevs_operational": 3, 00:10:04.724 "base_bdevs_list": [ 00:10:04.724 { 00:10:04.724 "name": "pt1", 00:10:04.724 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:04.724 "is_configured": true, 00:10:04.724 "data_offset": 2048, 00:10:04.724 "data_size": 63488 00:10:04.724 }, 00:10:04.724 { 00:10:04.724 "name": null, 00:10:04.724 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:04.724 "is_configured": false, 00:10:04.724 "data_offset": 0, 00:10:04.724 "data_size": 63488 00:10:04.724 }, 00:10:04.724 { 00:10:04.724 "name": null, 00:10:04.724 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:04.724 "is_configured": false, 00:10:04.724 "data_offset": 2048, 00:10:04.724 "data_size": 63488 00:10:04.724 } 00:10:04.724 ] 00:10:04.724 }' 00:10:04.724 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:04.724 15:15:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:05.291 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:10:05.291 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:05.291 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:05.291 15:15:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:05.291 15:15:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:05.291 [2024-11-19 15:15:55.456404] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:05.291 [2024-11-19 15:15:55.456504] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:05.291 [2024-11-19 15:15:55.456532] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:10:05.291 [2024-11-19 15:15:55.456544] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:05.291 [2024-11-19 15:15:55.457146] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:05.291 [2024-11-19 15:15:55.457174] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:05.291 [2024-11-19 15:15:55.457274] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:05.291 [2024-11-19 15:15:55.457308] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:05.291 pt2 00:10:05.291 15:15:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:05.291 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:05.291 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:05.291 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:05.291 15:15:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:05.291 15:15:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:05.291 [2024-11-19 15:15:55.468348] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:05.291 [2024-11-19 15:15:55.468406] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:05.291 [2024-11-19 15:15:55.468432] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:10:05.291 [2024-11-19 15:15:55.468443] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:05.291 [2024-11-19 15:15:55.468878] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:05.291 [2024-11-19 15:15:55.468897] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:05.291 [2024-11-19 15:15:55.468997] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:10:05.291 [2024-11-19 15:15:55.469019] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:05.291 [2024-11-19 15:15:55.469135] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:10:05.291 [2024-11-19 15:15:55.469145] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:10:05.291 [2024-11-19 15:15:55.469436] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:10:05.291 [2024-11-19 15:15:55.469560] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:10:05.291 [2024-11-19 15:15:55.469584] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:10:05.291 [2024-11-19 15:15:55.469722] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:05.291 pt3 00:10:05.291 15:15:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:05.291 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:05.291 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:05.291 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:10:05.291 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:05.291 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:05.291 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:05.291 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:05.291 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:05.291 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:05.292 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:05.292 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:05.292 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:05.292 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:05.292 15:15:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:05.292 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:05.292 15:15:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:05.292 15:15:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:05.292 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:05.292 "name": "raid_bdev1", 00:10:05.292 "uuid": "ea435670-ddbf-4a54-8fe3-340b4dab47c2", 00:10:05.292 "strip_size_kb": 64, 00:10:05.292 "state": "online", 00:10:05.292 "raid_level": "concat", 00:10:05.292 "superblock": true, 00:10:05.292 "num_base_bdevs": 3, 00:10:05.292 "num_base_bdevs_discovered": 3, 00:10:05.292 "num_base_bdevs_operational": 3, 00:10:05.292 "base_bdevs_list": [ 00:10:05.292 { 00:10:05.292 "name": "pt1", 00:10:05.292 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:05.292 "is_configured": true, 00:10:05.292 "data_offset": 2048, 00:10:05.292 "data_size": 63488 00:10:05.292 }, 00:10:05.292 { 00:10:05.292 "name": "pt2", 00:10:05.292 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:05.292 "is_configured": true, 00:10:05.292 "data_offset": 2048, 00:10:05.292 "data_size": 63488 00:10:05.292 }, 00:10:05.292 { 00:10:05.292 "name": "pt3", 00:10:05.292 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:05.292 "is_configured": true, 00:10:05.292 "data_offset": 2048, 00:10:05.292 "data_size": 63488 00:10:05.292 } 00:10:05.292 ] 00:10:05.292 }' 00:10:05.292 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:05.292 15:15:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:05.550 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:10:05.550 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:10:05.550 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:05.550 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:05.550 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:05.550 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:05.550 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:05.550 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:05.550 15:15:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:05.550 15:15:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:05.550 [2024-11-19 15:15:55.872259] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:05.810 15:15:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:05.810 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:05.810 "name": "raid_bdev1", 00:10:05.810 "aliases": [ 00:10:05.810 "ea435670-ddbf-4a54-8fe3-340b4dab47c2" 00:10:05.810 ], 00:10:05.810 "product_name": "Raid Volume", 00:10:05.810 "block_size": 512, 00:10:05.810 "num_blocks": 190464, 00:10:05.810 "uuid": "ea435670-ddbf-4a54-8fe3-340b4dab47c2", 00:10:05.810 "assigned_rate_limits": { 00:10:05.810 "rw_ios_per_sec": 0, 00:10:05.810 "rw_mbytes_per_sec": 0, 00:10:05.810 "r_mbytes_per_sec": 0, 00:10:05.810 "w_mbytes_per_sec": 0 00:10:05.810 }, 00:10:05.810 "claimed": false, 00:10:05.810 "zoned": false, 00:10:05.810 "supported_io_types": { 00:10:05.810 "read": true, 00:10:05.810 "write": true, 00:10:05.810 "unmap": true, 00:10:05.810 "flush": true, 00:10:05.810 "reset": true, 00:10:05.810 "nvme_admin": false, 00:10:05.810 "nvme_io": false, 00:10:05.810 "nvme_io_md": false, 00:10:05.810 "write_zeroes": true, 00:10:05.810 "zcopy": false, 00:10:05.810 "get_zone_info": false, 00:10:05.810 "zone_management": false, 00:10:05.810 "zone_append": false, 00:10:05.810 "compare": false, 00:10:05.810 "compare_and_write": false, 00:10:05.810 "abort": false, 00:10:05.810 "seek_hole": false, 00:10:05.810 "seek_data": false, 00:10:05.810 "copy": false, 00:10:05.810 "nvme_iov_md": false 00:10:05.810 }, 00:10:05.810 "memory_domains": [ 00:10:05.810 { 00:10:05.810 "dma_device_id": "system", 00:10:05.810 "dma_device_type": 1 00:10:05.810 }, 00:10:05.810 { 00:10:05.810 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:05.810 "dma_device_type": 2 00:10:05.810 }, 00:10:05.810 { 00:10:05.810 "dma_device_id": "system", 00:10:05.810 "dma_device_type": 1 00:10:05.810 }, 00:10:05.810 { 00:10:05.810 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:05.810 "dma_device_type": 2 00:10:05.810 }, 00:10:05.810 { 00:10:05.810 "dma_device_id": "system", 00:10:05.810 "dma_device_type": 1 00:10:05.810 }, 00:10:05.810 { 00:10:05.810 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:05.810 "dma_device_type": 2 00:10:05.810 } 00:10:05.810 ], 00:10:05.810 "driver_specific": { 00:10:05.810 "raid": { 00:10:05.810 "uuid": "ea435670-ddbf-4a54-8fe3-340b4dab47c2", 00:10:05.810 "strip_size_kb": 64, 00:10:05.810 "state": "online", 00:10:05.810 "raid_level": "concat", 00:10:05.810 "superblock": true, 00:10:05.810 "num_base_bdevs": 3, 00:10:05.810 "num_base_bdevs_discovered": 3, 00:10:05.810 "num_base_bdevs_operational": 3, 00:10:05.810 "base_bdevs_list": [ 00:10:05.810 { 00:10:05.810 "name": "pt1", 00:10:05.810 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:05.810 "is_configured": true, 00:10:05.810 "data_offset": 2048, 00:10:05.810 "data_size": 63488 00:10:05.810 }, 00:10:05.810 { 00:10:05.810 "name": "pt2", 00:10:05.810 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:05.810 "is_configured": true, 00:10:05.810 "data_offset": 2048, 00:10:05.810 "data_size": 63488 00:10:05.810 }, 00:10:05.810 { 00:10:05.810 "name": "pt3", 00:10:05.810 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:05.810 "is_configured": true, 00:10:05.810 "data_offset": 2048, 00:10:05.810 "data_size": 63488 00:10:05.810 } 00:10:05.810 ] 00:10:05.810 } 00:10:05.810 } 00:10:05.810 }' 00:10:05.811 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:05.811 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:10:05.811 pt2 00:10:05.811 pt3' 00:10:05.811 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:05.811 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:05.811 15:15:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:05.811 15:15:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:10:05.811 15:15:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:05.811 15:15:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:05.811 15:15:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:05.811 15:15:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:05.811 15:15:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:05.811 15:15:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:05.811 15:15:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:05.811 15:15:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:10:05.811 15:15:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:05.811 15:15:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:05.811 15:15:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:05.811 15:15:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:05.811 15:15:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:05.811 15:15:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:05.811 15:15:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:05.811 15:15:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:05.811 15:15:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:10:05.811 15:15:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:05.811 15:15:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:05.811 15:15:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:05.811 15:15:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:05.811 15:15:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:05.811 15:15:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:10:05.811 15:15:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:05.811 15:15:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:05.811 15:15:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:05.811 [2024-11-19 15:15:56.139650] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:06.071 15:15:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:06.071 15:15:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' ea435670-ddbf-4a54-8fe3-340b4dab47c2 '!=' ea435670-ddbf-4a54-8fe3-340b4dab47c2 ']' 00:10:06.071 15:15:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy concat 00:10:06.071 15:15:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:06.071 15:15:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:06.071 15:15:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 77950 00:10:06.071 15:15:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 77950 ']' 00:10:06.071 15:15:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 77950 00:10:06.071 15:15:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:10:06.071 15:15:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:06.071 15:15:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 77950 00:10:06.071 15:15:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:06.071 15:15:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:06.071 killing process with pid 77950 00:10:06.071 15:15:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 77950' 00:10:06.071 15:15:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 77950 00:10:06.071 [2024-11-19 15:15:56.206427] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:06.071 [2024-11-19 15:15:56.206555] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:06.071 15:15:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 77950 00:10:06.071 [2024-11-19 15:15:56.206638] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:06.071 [2024-11-19 15:15:56.206648] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:10:06.071 [2024-11-19 15:15:56.268919] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:06.332 15:15:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:10:06.332 00:10:06.332 real 0m3.984s 00:10:06.332 user 0m6.129s 00:10:06.332 sys 0m0.917s 00:10:06.332 15:15:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:06.332 15:15:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.332 ************************************ 00:10:06.332 END TEST raid_superblock_test 00:10:06.332 ************************************ 00:10:06.332 15:15:56 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test concat 3 read 00:10:06.332 15:15:56 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:10:06.332 15:15:56 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:06.332 15:15:56 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:06.332 ************************************ 00:10:06.332 START TEST raid_read_error_test 00:10:06.332 ************************************ 00:10:06.332 15:15:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test concat 3 read 00:10:06.332 15:15:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:10:06.332 15:15:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:10:06.332 15:15:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:10:06.332 15:15:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:10:06.332 15:15:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:06.332 15:15:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:10:06.332 15:15:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:06.332 15:15:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:06.332 15:15:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:10:06.332 15:15:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:06.332 15:15:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:06.332 15:15:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:10:06.332 15:15:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:06.332 15:15:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:06.332 15:15:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:10:06.332 15:15:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:10:06.591 15:15:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:10:06.591 15:15:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:10:06.591 15:15:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:10:06.591 15:15:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:10:06.591 15:15:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:10:06.591 15:15:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:10:06.591 15:15:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:10:06.591 15:15:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:10:06.591 15:15:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:10:06.591 15:15:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.NI7q5TAMSP 00:10:06.591 15:15:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=78192 00:10:06.591 15:15:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:10:06.591 15:15:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 78192 00:10:06.591 15:15:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 78192 ']' 00:10:06.591 15:15:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:06.591 15:15:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:06.592 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:06.592 15:15:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:06.592 15:15:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:06.592 15:15:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.592 [2024-11-19 15:15:56.763091] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:10:06.592 [2024-11-19 15:15:56.763228] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid78192 ] 00:10:06.592 [2024-11-19 15:15:56.917987] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:06.851 [2024-11-19 15:15:56.958184] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:06.851 [2024-11-19 15:15:57.034208] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:06.851 [2024-11-19 15:15:57.034247] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.420 BaseBdev1_malloc 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.420 true 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.420 [2024-11-19 15:15:57.632484] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:10:07.420 [2024-11-19 15:15:57.632555] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:07.420 [2024-11-19 15:15:57.632578] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:10:07.420 [2024-11-19 15:15:57.632588] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:07.420 [2024-11-19 15:15:57.635078] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:07.420 [2024-11-19 15:15:57.635111] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:07.420 BaseBdev1 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.420 BaseBdev2_malloc 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.420 true 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.420 [2024-11-19 15:15:57.679166] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:10:07.420 [2024-11-19 15:15:57.679212] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:07.420 [2024-11-19 15:15:57.679230] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:10:07.420 [2024-11-19 15:15:57.679246] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:07.420 [2024-11-19 15:15:57.681717] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:07.420 [2024-11-19 15:15:57.681754] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:10:07.420 BaseBdev2 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.420 BaseBdev3_malloc 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.420 true 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.420 [2024-11-19 15:15:57.725640] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:10:07.420 [2024-11-19 15:15:57.725686] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:07.420 [2024-11-19 15:15:57.725703] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:10:07.420 [2024-11-19 15:15:57.725712] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:07.420 [2024-11-19 15:15:57.728017] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:07.420 [2024-11-19 15:15:57.728047] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:10:07.420 BaseBdev3 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.420 [2024-11-19 15:15:57.737701] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:07.420 [2024-11-19 15:15:57.739797] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:07.420 [2024-11-19 15:15:57.739871] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:07.420 [2024-11-19 15:15:57.740062] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:10:07.420 [2024-11-19 15:15:57.740082] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:10:07.420 [2024-11-19 15:15:57.740339] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002bb0 00:10:07.420 [2024-11-19 15:15:57.740490] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:10:07.420 [2024-11-19 15:15:57.740506] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:10:07.420 [2024-11-19 15:15:57.740635] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:07.420 15:15:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:07.421 15:15:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:07.421 15:15:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:07.421 15:15:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:07.421 15:15:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:07.421 15:15:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:07.421 15:15:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.421 15:15:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.680 15:15:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.680 15:15:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:07.680 "name": "raid_bdev1", 00:10:07.680 "uuid": "6c654cd7-8a33-48fa-8b53-851fdbc7796e", 00:10:07.680 "strip_size_kb": 64, 00:10:07.680 "state": "online", 00:10:07.680 "raid_level": "concat", 00:10:07.680 "superblock": true, 00:10:07.680 "num_base_bdevs": 3, 00:10:07.680 "num_base_bdevs_discovered": 3, 00:10:07.680 "num_base_bdevs_operational": 3, 00:10:07.680 "base_bdevs_list": [ 00:10:07.680 { 00:10:07.680 "name": "BaseBdev1", 00:10:07.680 "uuid": "aba53542-8465-5291-9ebc-557592d85cab", 00:10:07.680 "is_configured": true, 00:10:07.680 "data_offset": 2048, 00:10:07.680 "data_size": 63488 00:10:07.680 }, 00:10:07.680 { 00:10:07.680 "name": "BaseBdev2", 00:10:07.680 "uuid": "3ca0e7ff-5335-595d-b8a9-ecd3d1e2a3dd", 00:10:07.680 "is_configured": true, 00:10:07.680 "data_offset": 2048, 00:10:07.680 "data_size": 63488 00:10:07.680 }, 00:10:07.680 { 00:10:07.681 "name": "BaseBdev3", 00:10:07.681 "uuid": "1ba186f3-d82f-59aa-b0b7-3799f620c8b9", 00:10:07.681 "is_configured": true, 00:10:07.681 "data_offset": 2048, 00:10:07.681 "data_size": 63488 00:10:07.681 } 00:10:07.681 ] 00:10:07.681 }' 00:10:07.681 15:15:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:07.681 15:15:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.945 15:15:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:10:07.945 15:15:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:10:07.945 [2024-11-19 15:15:58.261318] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002d50 00:10:08.886 15:15:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:10:08.886 15:15:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:08.886 15:15:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:08.886 15:15:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:08.886 15:15:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:10:08.886 15:15:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:10:08.886 15:15:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:10:08.886 15:15:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:10:08.886 15:15:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:08.886 15:15:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:08.886 15:15:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:08.886 15:15:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:08.886 15:15:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:08.886 15:15:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:08.886 15:15:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:08.886 15:15:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:08.886 15:15:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:08.886 15:15:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:08.886 15:15:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:08.886 15:15:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:08.886 15:15:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:08.886 15:15:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.145 15:15:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:09.145 "name": "raid_bdev1", 00:10:09.145 "uuid": "6c654cd7-8a33-48fa-8b53-851fdbc7796e", 00:10:09.145 "strip_size_kb": 64, 00:10:09.145 "state": "online", 00:10:09.145 "raid_level": "concat", 00:10:09.145 "superblock": true, 00:10:09.145 "num_base_bdevs": 3, 00:10:09.145 "num_base_bdevs_discovered": 3, 00:10:09.145 "num_base_bdevs_operational": 3, 00:10:09.145 "base_bdevs_list": [ 00:10:09.145 { 00:10:09.145 "name": "BaseBdev1", 00:10:09.145 "uuid": "aba53542-8465-5291-9ebc-557592d85cab", 00:10:09.145 "is_configured": true, 00:10:09.145 "data_offset": 2048, 00:10:09.145 "data_size": 63488 00:10:09.145 }, 00:10:09.145 { 00:10:09.145 "name": "BaseBdev2", 00:10:09.145 "uuid": "3ca0e7ff-5335-595d-b8a9-ecd3d1e2a3dd", 00:10:09.145 "is_configured": true, 00:10:09.145 "data_offset": 2048, 00:10:09.145 "data_size": 63488 00:10:09.145 }, 00:10:09.145 { 00:10:09.145 "name": "BaseBdev3", 00:10:09.145 "uuid": "1ba186f3-d82f-59aa-b0b7-3799f620c8b9", 00:10:09.145 "is_configured": true, 00:10:09.145 "data_offset": 2048, 00:10:09.145 "data_size": 63488 00:10:09.145 } 00:10:09.145 ] 00:10:09.145 }' 00:10:09.145 15:15:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:09.145 15:15:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.404 15:15:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:09.404 15:15:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.404 15:15:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.404 [2024-11-19 15:15:59.569872] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:09.404 [2024-11-19 15:15:59.569928] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:09.404 [2024-11-19 15:15:59.572621] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:09.404 [2024-11-19 15:15:59.572692] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:09.404 [2024-11-19 15:15:59.572734] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:09.404 [2024-11-19 15:15:59.572748] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:10:09.404 { 00:10:09.404 "results": [ 00:10:09.404 { 00:10:09.404 "job": "raid_bdev1", 00:10:09.404 "core_mask": "0x1", 00:10:09.404 "workload": "randrw", 00:10:09.404 "percentage": 50, 00:10:09.404 "status": "finished", 00:10:09.404 "queue_depth": 1, 00:10:09.404 "io_size": 131072, 00:10:09.404 "runtime": 1.309051, 00:10:09.404 "iops": 14030.01105380921, 00:10:09.404 "mibps": 1753.7513817261513, 00:10:09.404 "io_failed": 1, 00:10:09.404 "io_timeout": 0, 00:10:09.404 "avg_latency_us": 99.9454561924355, 00:10:09.404 "min_latency_us": 25.6, 00:10:09.404 "max_latency_us": 1402.2986899563318 00:10:09.404 } 00:10:09.404 ], 00:10:09.404 "core_count": 1 00:10:09.404 } 00:10:09.404 15:15:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.404 15:15:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 78192 00:10:09.404 15:15:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 78192 ']' 00:10:09.404 15:15:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 78192 00:10:09.404 15:15:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:10:09.404 15:15:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:09.404 15:15:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 78192 00:10:09.404 15:15:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:09.404 15:15:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:09.404 killing process with pid 78192 00:10:09.404 15:15:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 78192' 00:10:09.404 15:15:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 78192 00:10:09.405 [2024-11-19 15:15:59.618620] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:09.405 15:15:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 78192 00:10:09.405 [2024-11-19 15:15:59.668296] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:09.663 15:15:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:10:09.663 15:15:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.NI7q5TAMSP 00:10:09.663 15:15:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:10:09.663 15:15:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.76 00:10:09.663 15:15:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:10:09.663 15:15:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:09.663 15:15:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:09.663 15:15:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.76 != \0\.\0\0 ]] 00:10:09.663 00:10:09.663 real 0m3.338s 00:10:09.663 user 0m4.076s 00:10:09.663 sys 0m0.608s 00:10:09.663 15:15:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:09.663 15:15:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.663 ************************************ 00:10:09.663 END TEST raid_read_error_test 00:10:09.663 ************************************ 00:10:09.922 15:16:00 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test concat 3 write 00:10:09.922 15:16:00 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:10:09.922 15:16:00 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:09.922 15:16:00 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:09.922 ************************************ 00:10:09.922 START TEST raid_write_error_test 00:10:09.922 ************************************ 00:10:09.922 15:16:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test concat 3 write 00:10:09.922 15:16:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:10:09.922 15:16:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:10:09.922 15:16:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:10:09.922 15:16:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:10:09.922 15:16:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:09.922 15:16:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:10:09.922 15:16:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:09.922 15:16:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:09.922 15:16:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:10:09.922 15:16:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:09.922 15:16:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:09.922 15:16:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:10:09.922 15:16:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:09.922 15:16:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:09.922 15:16:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:10:09.922 15:16:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:10:09.922 15:16:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:10:09.922 15:16:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:10:09.922 15:16:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:10:09.922 15:16:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:10:09.922 15:16:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:10:09.922 15:16:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:10:09.922 15:16:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:10:09.922 15:16:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:10:09.922 15:16:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:10:09.922 15:16:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.XCyaBudRxt 00:10:09.922 15:16:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=78321 00:10:09.922 15:16:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:10:09.922 15:16:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 78321 00:10:09.922 15:16:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 78321 ']' 00:10:09.922 15:16:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:09.922 15:16:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:09.922 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:09.922 15:16:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:09.922 15:16:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:09.923 15:16:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.923 [2024-11-19 15:16:00.175996] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:10:09.923 [2024-11-19 15:16:00.176143] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid78321 ] 00:10:10.181 [2024-11-19 15:16:00.333951] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:10.181 [2024-11-19 15:16:00.374811] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:10.181 [2024-11-19 15:16:00.451216] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:10.181 [2024-11-19 15:16:00.451262] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:10.748 15:16:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:10.748 15:16:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:10:10.748 15:16:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:10.748 15:16:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:10:10.748 15:16:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:10.748 15:16:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.748 BaseBdev1_malloc 00:10:10.748 15:16:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:10.748 15:16:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:10:10.748 15:16:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:10.748 15:16:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.748 true 00:10:10.749 15:16:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:10.749 15:16:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:10:10.749 15:16:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:10.749 15:16:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.749 [2024-11-19 15:16:01.061143] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:10:10.749 [2024-11-19 15:16:01.061217] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:10.749 [2024-11-19 15:16:01.061259] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:10:10.749 [2024-11-19 15:16:01.061274] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:10.749 [2024-11-19 15:16:01.063824] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:10.749 [2024-11-19 15:16:01.063859] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:10.749 BaseBdev1 00:10:10.749 15:16:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:10.749 15:16:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:10.749 15:16:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:10:10.749 15:16:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:10.749 15:16:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.008 BaseBdev2_malloc 00:10:11.008 15:16:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.008 15:16:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:10:11.008 15:16:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.008 15:16:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.008 true 00:10:11.008 15:16:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.008 15:16:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:10:11.008 15:16:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.008 15:16:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.008 [2024-11-19 15:16:01.107901] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:10:11.008 [2024-11-19 15:16:01.107956] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:11.008 [2024-11-19 15:16:01.108002] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:10:11.008 [2024-11-19 15:16:01.108022] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:11.008 [2024-11-19 15:16:01.110497] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:11.008 [2024-11-19 15:16:01.110534] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:10:11.008 BaseBdev2 00:10:11.008 15:16:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.008 15:16:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:11.008 15:16:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:10:11.008 15:16:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.008 15:16:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.008 BaseBdev3_malloc 00:10:11.008 15:16:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.008 15:16:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:10:11.008 15:16:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.008 15:16:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.008 true 00:10:11.008 15:16:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.008 15:16:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:10:11.008 15:16:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.008 15:16:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.008 [2024-11-19 15:16:01.154514] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:10:11.008 [2024-11-19 15:16:01.154569] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:11.008 [2024-11-19 15:16:01.154588] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:10:11.008 [2024-11-19 15:16:01.154597] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:11.008 [2024-11-19 15:16:01.157097] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:11.008 [2024-11-19 15:16:01.157131] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:10:11.008 BaseBdev3 00:10:11.008 15:16:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.008 15:16:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:10:11.008 15:16:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.008 15:16:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.008 [2024-11-19 15:16:01.166586] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:11.008 [2024-11-19 15:16:01.168802] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:11.008 [2024-11-19 15:16:01.168879] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:11.008 [2024-11-19 15:16:01.169133] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:10:11.008 [2024-11-19 15:16:01.169167] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:10:11.008 [2024-11-19 15:16:01.169448] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002bb0 00:10:11.008 [2024-11-19 15:16:01.169595] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:10:11.008 [2024-11-19 15:16:01.169609] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:10:11.008 [2024-11-19 15:16:01.169756] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:11.008 15:16:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.008 15:16:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:10:11.008 15:16:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:11.008 15:16:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:11.008 15:16:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:11.008 15:16:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:11.008 15:16:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:11.008 15:16:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:11.009 15:16:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:11.009 15:16:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:11.009 15:16:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:11.009 15:16:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:11.009 15:16:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:11.009 15:16:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.009 15:16:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.009 15:16:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.009 15:16:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:11.009 "name": "raid_bdev1", 00:10:11.009 "uuid": "6529f0fe-7bb9-4ef7-a521-7fa78f22efd7", 00:10:11.009 "strip_size_kb": 64, 00:10:11.009 "state": "online", 00:10:11.009 "raid_level": "concat", 00:10:11.009 "superblock": true, 00:10:11.009 "num_base_bdevs": 3, 00:10:11.009 "num_base_bdevs_discovered": 3, 00:10:11.009 "num_base_bdevs_operational": 3, 00:10:11.009 "base_bdevs_list": [ 00:10:11.009 { 00:10:11.009 "name": "BaseBdev1", 00:10:11.009 "uuid": "8ea2adf3-63df-5719-82e8-8ee2e3067dbb", 00:10:11.009 "is_configured": true, 00:10:11.009 "data_offset": 2048, 00:10:11.009 "data_size": 63488 00:10:11.009 }, 00:10:11.009 { 00:10:11.009 "name": "BaseBdev2", 00:10:11.009 "uuid": "6352c248-82b7-57cb-b2cd-4ac7f2b2e32a", 00:10:11.009 "is_configured": true, 00:10:11.009 "data_offset": 2048, 00:10:11.009 "data_size": 63488 00:10:11.009 }, 00:10:11.009 { 00:10:11.009 "name": "BaseBdev3", 00:10:11.009 "uuid": "2980a34c-bc23-592b-a1fa-6cbfe9865718", 00:10:11.009 "is_configured": true, 00:10:11.009 "data_offset": 2048, 00:10:11.009 "data_size": 63488 00:10:11.009 } 00:10:11.009 ] 00:10:11.009 }' 00:10:11.009 15:16:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:11.009 15:16:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.267 15:16:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:10:11.267 15:16:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:10:11.527 [2024-11-19 15:16:01.650326] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002d50 00:10:12.464 15:16:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:10:12.464 15:16:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:12.464 15:16:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.464 15:16:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:12.464 15:16:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:10:12.464 15:16:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:10:12.464 15:16:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:10:12.464 15:16:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:10:12.464 15:16:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:12.464 15:16:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:12.464 15:16:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:12.464 15:16:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:12.464 15:16:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:12.464 15:16:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:12.464 15:16:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:12.464 15:16:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:12.464 15:16:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:12.464 15:16:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:12.464 15:16:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:12.464 15:16:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:12.464 15:16:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.464 15:16:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:12.464 15:16:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:12.464 "name": "raid_bdev1", 00:10:12.464 "uuid": "6529f0fe-7bb9-4ef7-a521-7fa78f22efd7", 00:10:12.464 "strip_size_kb": 64, 00:10:12.464 "state": "online", 00:10:12.464 "raid_level": "concat", 00:10:12.464 "superblock": true, 00:10:12.464 "num_base_bdevs": 3, 00:10:12.464 "num_base_bdevs_discovered": 3, 00:10:12.464 "num_base_bdevs_operational": 3, 00:10:12.464 "base_bdevs_list": [ 00:10:12.464 { 00:10:12.464 "name": "BaseBdev1", 00:10:12.464 "uuid": "8ea2adf3-63df-5719-82e8-8ee2e3067dbb", 00:10:12.464 "is_configured": true, 00:10:12.464 "data_offset": 2048, 00:10:12.464 "data_size": 63488 00:10:12.464 }, 00:10:12.464 { 00:10:12.464 "name": "BaseBdev2", 00:10:12.464 "uuid": "6352c248-82b7-57cb-b2cd-4ac7f2b2e32a", 00:10:12.464 "is_configured": true, 00:10:12.464 "data_offset": 2048, 00:10:12.464 "data_size": 63488 00:10:12.464 }, 00:10:12.464 { 00:10:12.464 "name": "BaseBdev3", 00:10:12.465 "uuid": "2980a34c-bc23-592b-a1fa-6cbfe9865718", 00:10:12.465 "is_configured": true, 00:10:12.465 "data_offset": 2048, 00:10:12.465 "data_size": 63488 00:10:12.465 } 00:10:12.465 ] 00:10:12.465 }' 00:10:12.465 15:16:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:12.465 15:16:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.723 15:16:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:12.724 15:16:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:12.724 15:16:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.724 [2024-11-19 15:16:03.014936] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:12.724 [2024-11-19 15:16:03.015005] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:12.724 [2024-11-19 15:16:03.017705] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:12.724 [2024-11-19 15:16:03.017770] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:12.724 [2024-11-19 15:16:03.017814] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:12.724 [2024-11-19 15:16:03.017833] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:10:12.724 { 00:10:12.724 "results": [ 00:10:12.724 { 00:10:12.724 "job": "raid_bdev1", 00:10:12.724 "core_mask": "0x1", 00:10:12.724 "workload": "randrw", 00:10:12.724 "percentage": 50, 00:10:12.724 "status": "finished", 00:10:12.724 "queue_depth": 1, 00:10:12.724 "io_size": 131072, 00:10:12.724 "runtime": 1.365052, 00:10:12.724 "iops": 14303.484409385137, 00:10:12.724 "mibps": 1787.9355511731421, 00:10:12.724 "io_failed": 1, 00:10:12.724 "io_timeout": 0, 00:10:12.724 "avg_latency_us": 98.1184124895392, 00:10:12.724 "min_latency_us": 27.053275109170304, 00:10:12.724 "max_latency_us": 1430.9170305676855 00:10:12.724 } 00:10:12.724 ], 00:10:12.724 "core_count": 1 00:10:12.724 } 00:10:12.724 15:16:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:12.724 15:16:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 78321 00:10:12.724 15:16:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 78321 ']' 00:10:12.724 15:16:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 78321 00:10:12.724 15:16:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:10:12.724 15:16:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:12.724 15:16:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 78321 00:10:12.724 15:16:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:12.724 15:16:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:12.724 15:16:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 78321' 00:10:12.724 killing process with pid 78321 00:10:12.724 15:16:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 78321 00:10:12.724 [2024-11-19 15:16:03.052167] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:12.724 15:16:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 78321 00:10:12.982 [2024-11-19 15:16:03.101764] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:13.241 15:16:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:10:13.241 15:16:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.XCyaBudRxt 00:10:13.241 15:16:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:10:13.241 15:16:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.73 00:10:13.241 15:16:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:10:13.241 15:16:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:13.241 15:16:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:13.241 15:16:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.73 != \0\.\0\0 ]] 00:10:13.241 00:10:13.241 real 0m3.355s 00:10:13.241 user 0m4.129s 00:10:13.241 sys 0m0.589s 00:10:13.241 15:16:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:13.241 15:16:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.241 ************************************ 00:10:13.241 END TEST raid_write_error_test 00:10:13.241 ************************************ 00:10:13.241 15:16:03 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:10:13.241 15:16:03 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid1 3 false 00:10:13.241 15:16:03 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:10:13.241 15:16:03 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:13.241 15:16:03 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:13.241 ************************************ 00:10:13.241 START TEST raid_state_function_test 00:10:13.241 ************************************ 00:10:13.241 15:16:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 3 false 00:10:13.241 15:16:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:10:13.241 15:16:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:10:13.241 15:16:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:10:13.241 15:16:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:10:13.241 15:16:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:10:13.241 15:16:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:13.241 15:16:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:10:13.241 15:16:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:13.241 15:16:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:13.241 15:16:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:10:13.241 15:16:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:13.241 15:16:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:13.241 15:16:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:10:13.241 15:16:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:13.241 15:16:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:13.241 15:16:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:10:13.242 15:16:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:10:13.242 15:16:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:10:13.242 15:16:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:10:13.242 15:16:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:10:13.242 15:16:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:10:13.242 15:16:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:10:13.242 15:16:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:10:13.242 15:16:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:10:13.242 15:16:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:10:13.242 15:16:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=78454 00:10:13.242 15:16:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:10:13.242 15:16:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 78454' 00:10:13.242 Process raid pid: 78454 00:10:13.242 15:16:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 78454 00:10:13.242 15:16:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 78454 ']' 00:10:13.242 15:16:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:13.242 15:16:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:13.242 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:13.242 15:16:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:13.242 15:16:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:13.242 15:16:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.500 [2024-11-19 15:16:03.588053] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:10:13.500 [2024-11-19 15:16:03.588202] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:13.500 [2024-11-19 15:16:03.747296] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:13.500 [2024-11-19 15:16:03.788493] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:13.758 [2024-11-19 15:16:03.866895] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:13.758 [2024-11-19 15:16:03.866944] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:14.325 15:16:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:14.325 15:16:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:10:14.325 15:16:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:10:14.325 15:16:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:14.325 15:16:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.325 [2024-11-19 15:16:04.454358] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:14.325 [2024-11-19 15:16:04.454438] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:14.325 [2024-11-19 15:16:04.454456] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:14.325 [2024-11-19 15:16:04.454470] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:14.325 [2024-11-19 15:16:04.454476] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:14.325 [2024-11-19 15:16:04.454490] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:14.325 15:16:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:14.325 15:16:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:10:14.325 15:16:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:14.325 15:16:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:14.325 15:16:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:14.325 15:16:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:14.325 15:16:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:14.325 15:16:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:14.325 15:16:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:14.325 15:16:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:14.325 15:16:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:14.325 15:16:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:14.325 15:16:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:14.325 15:16:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:14.325 15:16:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.325 15:16:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:14.325 15:16:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:14.325 "name": "Existed_Raid", 00:10:14.325 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:14.325 "strip_size_kb": 0, 00:10:14.325 "state": "configuring", 00:10:14.325 "raid_level": "raid1", 00:10:14.325 "superblock": false, 00:10:14.325 "num_base_bdevs": 3, 00:10:14.325 "num_base_bdevs_discovered": 0, 00:10:14.325 "num_base_bdevs_operational": 3, 00:10:14.325 "base_bdevs_list": [ 00:10:14.325 { 00:10:14.325 "name": "BaseBdev1", 00:10:14.325 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:14.325 "is_configured": false, 00:10:14.325 "data_offset": 0, 00:10:14.325 "data_size": 0 00:10:14.325 }, 00:10:14.325 { 00:10:14.325 "name": "BaseBdev2", 00:10:14.325 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:14.325 "is_configured": false, 00:10:14.325 "data_offset": 0, 00:10:14.325 "data_size": 0 00:10:14.325 }, 00:10:14.325 { 00:10:14.325 "name": "BaseBdev3", 00:10:14.325 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:14.325 "is_configured": false, 00:10:14.325 "data_offset": 0, 00:10:14.325 "data_size": 0 00:10:14.325 } 00:10:14.325 ] 00:10:14.325 }' 00:10:14.325 15:16:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:14.325 15:16:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.584 15:16:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:14.584 15:16:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:14.584 15:16:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.584 [2024-11-19 15:16:04.869689] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:14.584 [2024-11-19 15:16:04.869753] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:10:14.584 15:16:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:14.584 15:16:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:10:14.584 15:16:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:14.584 15:16:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.584 [2024-11-19 15:16:04.881639] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:14.584 [2024-11-19 15:16:04.881691] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:14.584 [2024-11-19 15:16:04.881700] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:14.584 [2024-11-19 15:16:04.881710] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:14.584 [2024-11-19 15:16:04.881716] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:14.584 [2024-11-19 15:16:04.881726] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:14.584 15:16:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:14.584 15:16:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:14.584 15:16:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:14.584 15:16:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.584 [2024-11-19 15:16:04.908664] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:14.584 BaseBdev1 00:10:14.584 15:16:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:14.584 15:16:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:10:14.584 15:16:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:10:14.584 15:16:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:14.584 15:16:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:14.584 15:16:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:14.584 15:16:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:14.584 15:16:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:14.584 15:16:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:14.584 15:16:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.584 15:16:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:14.584 15:16:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:14.584 15:16:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:14.843 15:16:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.843 [ 00:10:14.843 { 00:10:14.843 "name": "BaseBdev1", 00:10:14.843 "aliases": [ 00:10:14.843 "ffaf53e2-022f-4529-a0da-aafc64760c9c" 00:10:14.843 ], 00:10:14.843 "product_name": "Malloc disk", 00:10:14.844 "block_size": 512, 00:10:14.844 "num_blocks": 65536, 00:10:14.844 "uuid": "ffaf53e2-022f-4529-a0da-aafc64760c9c", 00:10:14.844 "assigned_rate_limits": { 00:10:14.844 "rw_ios_per_sec": 0, 00:10:14.844 "rw_mbytes_per_sec": 0, 00:10:14.844 "r_mbytes_per_sec": 0, 00:10:14.844 "w_mbytes_per_sec": 0 00:10:14.844 }, 00:10:14.844 "claimed": true, 00:10:14.844 "claim_type": "exclusive_write", 00:10:14.844 "zoned": false, 00:10:14.844 "supported_io_types": { 00:10:14.844 "read": true, 00:10:14.844 "write": true, 00:10:14.844 "unmap": true, 00:10:14.844 "flush": true, 00:10:14.844 "reset": true, 00:10:14.844 "nvme_admin": false, 00:10:14.844 "nvme_io": false, 00:10:14.844 "nvme_io_md": false, 00:10:14.844 "write_zeroes": true, 00:10:14.844 "zcopy": true, 00:10:14.844 "get_zone_info": false, 00:10:14.844 "zone_management": false, 00:10:14.844 "zone_append": false, 00:10:14.844 "compare": false, 00:10:14.844 "compare_and_write": false, 00:10:14.844 "abort": true, 00:10:14.844 "seek_hole": false, 00:10:14.844 "seek_data": false, 00:10:14.844 "copy": true, 00:10:14.844 "nvme_iov_md": false 00:10:14.844 }, 00:10:14.844 "memory_domains": [ 00:10:14.844 { 00:10:14.844 "dma_device_id": "system", 00:10:14.844 "dma_device_type": 1 00:10:14.844 }, 00:10:14.844 { 00:10:14.844 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:14.844 "dma_device_type": 2 00:10:14.844 } 00:10:14.844 ], 00:10:14.844 "driver_specific": {} 00:10:14.844 } 00:10:14.844 ] 00:10:14.844 15:16:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:14.844 15:16:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:14.844 15:16:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:10:14.844 15:16:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:14.844 15:16:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:14.844 15:16:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:14.844 15:16:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:14.844 15:16:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:14.844 15:16:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:14.844 15:16:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:14.844 15:16:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:14.844 15:16:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:14.844 15:16:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:14.844 15:16:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:14.844 15:16:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:14.844 15:16:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.844 15:16:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:14.844 15:16:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:14.844 "name": "Existed_Raid", 00:10:14.844 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:14.844 "strip_size_kb": 0, 00:10:14.844 "state": "configuring", 00:10:14.844 "raid_level": "raid1", 00:10:14.844 "superblock": false, 00:10:14.844 "num_base_bdevs": 3, 00:10:14.844 "num_base_bdevs_discovered": 1, 00:10:14.844 "num_base_bdevs_operational": 3, 00:10:14.844 "base_bdevs_list": [ 00:10:14.844 { 00:10:14.844 "name": "BaseBdev1", 00:10:14.844 "uuid": "ffaf53e2-022f-4529-a0da-aafc64760c9c", 00:10:14.844 "is_configured": true, 00:10:14.844 "data_offset": 0, 00:10:14.844 "data_size": 65536 00:10:14.844 }, 00:10:14.844 { 00:10:14.844 "name": "BaseBdev2", 00:10:14.844 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:14.844 "is_configured": false, 00:10:14.844 "data_offset": 0, 00:10:14.844 "data_size": 0 00:10:14.844 }, 00:10:14.844 { 00:10:14.844 "name": "BaseBdev3", 00:10:14.844 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:14.844 "is_configured": false, 00:10:14.844 "data_offset": 0, 00:10:14.844 "data_size": 0 00:10:14.844 } 00:10:14.844 ] 00:10:14.844 }' 00:10:14.844 15:16:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:14.844 15:16:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.103 15:16:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:15.103 15:16:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:15.104 15:16:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.104 [2024-11-19 15:16:05.360054] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:15.104 [2024-11-19 15:16:05.360140] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:10:15.104 15:16:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:15.104 15:16:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:10:15.104 15:16:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:15.104 15:16:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.104 [2024-11-19 15:16:05.368019] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:15.104 [2024-11-19 15:16:05.370342] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:15.104 [2024-11-19 15:16:05.370387] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:15.104 [2024-11-19 15:16:05.370397] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:15.104 [2024-11-19 15:16:05.370409] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:15.104 15:16:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:15.104 15:16:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:10:15.104 15:16:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:15.104 15:16:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:10:15.104 15:16:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:15.104 15:16:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:15.104 15:16:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:15.104 15:16:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:15.104 15:16:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:15.104 15:16:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:15.104 15:16:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:15.104 15:16:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:15.104 15:16:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:15.104 15:16:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:15.104 15:16:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:15.104 15:16:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:15.104 15:16:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.104 15:16:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:15.104 15:16:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:15.104 "name": "Existed_Raid", 00:10:15.104 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:15.104 "strip_size_kb": 0, 00:10:15.104 "state": "configuring", 00:10:15.104 "raid_level": "raid1", 00:10:15.104 "superblock": false, 00:10:15.104 "num_base_bdevs": 3, 00:10:15.104 "num_base_bdevs_discovered": 1, 00:10:15.104 "num_base_bdevs_operational": 3, 00:10:15.104 "base_bdevs_list": [ 00:10:15.104 { 00:10:15.104 "name": "BaseBdev1", 00:10:15.104 "uuid": "ffaf53e2-022f-4529-a0da-aafc64760c9c", 00:10:15.104 "is_configured": true, 00:10:15.104 "data_offset": 0, 00:10:15.104 "data_size": 65536 00:10:15.104 }, 00:10:15.104 { 00:10:15.104 "name": "BaseBdev2", 00:10:15.104 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:15.104 "is_configured": false, 00:10:15.104 "data_offset": 0, 00:10:15.104 "data_size": 0 00:10:15.104 }, 00:10:15.104 { 00:10:15.104 "name": "BaseBdev3", 00:10:15.104 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:15.104 "is_configured": false, 00:10:15.104 "data_offset": 0, 00:10:15.104 "data_size": 0 00:10:15.104 } 00:10:15.104 ] 00:10:15.104 }' 00:10:15.104 15:16:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:15.104 15:16:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.672 15:16:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:15.672 15:16:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:15.672 15:16:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.672 [2024-11-19 15:16:05.823936] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:15.672 BaseBdev2 00:10:15.672 15:16:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:15.672 15:16:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:10:15.672 15:16:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:10:15.672 15:16:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:15.672 15:16:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:15.672 15:16:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:15.672 15:16:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:15.672 15:16:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:15.672 15:16:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:15.672 15:16:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.672 15:16:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:15.672 15:16:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:15.672 15:16:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:15.672 15:16:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.672 [ 00:10:15.672 { 00:10:15.672 "name": "BaseBdev2", 00:10:15.672 "aliases": [ 00:10:15.672 "fc8164fc-3b1e-4a0d-bcb3-257ced4f6e9c" 00:10:15.672 ], 00:10:15.672 "product_name": "Malloc disk", 00:10:15.672 "block_size": 512, 00:10:15.672 "num_blocks": 65536, 00:10:15.672 "uuid": "fc8164fc-3b1e-4a0d-bcb3-257ced4f6e9c", 00:10:15.672 "assigned_rate_limits": { 00:10:15.672 "rw_ios_per_sec": 0, 00:10:15.672 "rw_mbytes_per_sec": 0, 00:10:15.672 "r_mbytes_per_sec": 0, 00:10:15.672 "w_mbytes_per_sec": 0 00:10:15.672 }, 00:10:15.672 "claimed": true, 00:10:15.672 "claim_type": "exclusive_write", 00:10:15.672 "zoned": false, 00:10:15.672 "supported_io_types": { 00:10:15.672 "read": true, 00:10:15.672 "write": true, 00:10:15.672 "unmap": true, 00:10:15.672 "flush": true, 00:10:15.672 "reset": true, 00:10:15.672 "nvme_admin": false, 00:10:15.672 "nvme_io": false, 00:10:15.672 "nvme_io_md": false, 00:10:15.672 "write_zeroes": true, 00:10:15.672 "zcopy": true, 00:10:15.672 "get_zone_info": false, 00:10:15.672 "zone_management": false, 00:10:15.672 "zone_append": false, 00:10:15.672 "compare": false, 00:10:15.672 "compare_and_write": false, 00:10:15.672 "abort": true, 00:10:15.672 "seek_hole": false, 00:10:15.672 "seek_data": false, 00:10:15.672 "copy": true, 00:10:15.672 "nvme_iov_md": false 00:10:15.672 }, 00:10:15.672 "memory_domains": [ 00:10:15.672 { 00:10:15.672 "dma_device_id": "system", 00:10:15.672 "dma_device_type": 1 00:10:15.672 }, 00:10:15.672 { 00:10:15.672 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:15.672 "dma_device_type": 2 00:10:15.672 } 00:10:15.672 ], 00:10:15.672 "driver_specific": {} 00:10:15.672 } 00:10:15.672 ] 00:10:15.672 15:16:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:15.672 15:16:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:15.672 15:16:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:15.672 15:16:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:15.672 15:16:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:10:15.672 15:16:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:15.672 15:16:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:15.672 15:16:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:15.672 15:16:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:15.672 15:16:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:15.672 15:16:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:15.672 15:16:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:15.672 15:16:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:15.672 15:16:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:15.672 15:16:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:15.673 15:16:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:15.673 15:16:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:15.673 15:16:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.673 15:16:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:15.673 15:16:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:15.673 "name": "Existed_Raid", 00:10:15.673 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:15.673 "strip_size_kb": 0, 00:10:15.673 "state": "configuring", 00:10:15.673 "raid_level": "raid1", 00:10:15.673 "superblock": false, 00:10:15.673 "num_base_bdevs": 3, 00:10:15.673 "num_base_bdevs_discovered": 2, 00:10:15.673 "num_base_bdevs_operational": 3, 00:10:15.673 "base_bdevs_list": [ 00:10:15.673 { 00:10:15.673 "name": "BaseBdev1", 00:10:15.673 "uuid": "ffaf53e2-022f-4529-a0da-aafc64760c9c", 00:10:15.673 "is_configured": true, 00:10:15.673 "data_offset": 0, 00:10:15.673 "data_size": 65536 00:10:15.673 }, 00:10:15.673 { 00:10:15.673 "name": "BaseBdev2", 00:10:15.673 "uuid": "fc8164fc-3b1e-4a0d-bcb3-257ced4f6e9c", 00:10:15.673 "is_configured": true, 00:10:15.673 "data_offset": 0, 00:10:15.673 "data_size": 65536 00:10:15.673 }, 00:10:15.673 { 00:10:15.673 "name": "BaseBdev3", 00:10:15.673 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:15.673 "is_configured": false, 00:10:15.673 "data_offset": 0, 00:10:15.673 "data_size": 0 00:10:15.673 } 00:10:15.673 ] 00:10:15.673 }' 00:10:15.673 15:16:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:15.673 15:16:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.241 15:16:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:16.241 15:16:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:16.241 15:16:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.241 [2024-11-19 15:16:06.327182] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:16.241 [2024-11-19 15:16:06.327242] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:10:16.241 [2024-11-19 15:16:06.327258] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:10:16.241 [2024-11-19 15:16:06.327614] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:10:16.241 [2024-11-19 15:16:06.327793] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:10:16.241 [2024-11-19 15:16:06.327812] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:10:16.241 [2024-11-19 15:16:06.328080] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:16.241 BaseBdev3 00:10:16.241 15:16:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:16.241 15:16:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:10:16.241 15:16:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:10:16.241 15:16:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:16.241 15:16:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:16.241 15:16:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:16.241 15:16:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:16.241 15:16:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:16.241 15:16:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:16.242 15:16:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.242 15:16:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:16.242 15:16:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:16.242 15:16:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:16.242 15:16:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.242 [ 00:10:16.242 { 00:10:16.242 "name": "BaseBdev3", 00:10:16.242 "aliases": [ 00:10:16.242 "ae268b03-50be-48a4-85b4-3f732353a86b" 00:10:16.242 ], 00:10:16.242 "product_name": "Malloc disk", 00:10:16.242 "block_size": 512, 00:10:16.242 "num_blocks": 65536, 00:10:16.242 "uuid": "ae268b03-50be-48a4-85b4-3f732353a86b", 00:10:16.242 "assigned_rate_limits": { 00:10:16.242 "rw_ios_per_sec": 0, 00:10:16.242 "rw_mbytes_per_sec": 0, 00:10:16.242 "r_mbytes_per_sec": 0, 00:10:16.242 "w_mbytes_per_sec": 0 00:10:16.242 }, 00:10:16.242 "claimed": true, 00:10:16.242 "claim_type": "exclusive_write", 00:10:16.242 "zoned": false, 00:10:16.242 "supported_io_types": { 00:10:16.242 "read": true, 00:10:16.242 "write": true, 00:10:16.242 "unmap": true, 00:10:16.242 "flush": true, 00:10:16.242 "reset": true, 00:10:16.242 "nvme_admin": false, 00:10:16.242 "nvme_io": false, 00:10:16.242 "nvme_io_md": false, 00:10:16.242 "write_zeroes": true, 00:10:16.242 "zcopy": true, 00:10:16.242 "get_zone_info": false, 00:10:16.242 "zone_management": false, 00:10:16.242 "zone_append": false, 00:10:16.242 "compare": false, 00:10:16.242 "compare_and_write": false, 00:10:16.242 "abort": true, 00:10:16.242 "seek_hole": false, 00:10:16.242 "seek_data": false, 00:10:16.242 "copy": true, 00:10:16.242 "nvme_iov_md": false 00:10:16.242 }, 00:10:16.242 "memory_domains": [ 00:10:16.242 { 00:10:16.242 "dma_device_id": "system", 00:10:16.242 "dma_device_type": 1 00:10:16.242 }, 00:10:16.242 { 00:10:16.242 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:16.242 "dma_device_type": 2 00:10:16.242 } 00:10:16.242 ], 00:10:16.242 "driver_specific": {} 00:10:16.242 } 00:10:16.242 ] 00:10:16.242 15:16:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:16.242 15:16:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:16.242 15:16:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:16.242 15:16:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:16.242 15:16:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:10:16.242 15:16:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:16.242 15:16:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:16.242 15:16:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:16.242 15:16:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:16.242 15:16:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:16.242 15:16:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:16.242 15:16:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:16.242 15:16:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:16.242 15:16:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:16.242 15:16:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:16.242 15:16:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:16.242 15:16:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:16.242 15:16:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.242 15:16:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:16.242 15:16:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:16.242 "name": "Existed_Raid", 00:10:16.242 "uuid": "35b7e949-0cc9-4592-ab59-bca19d24befc", 00:10:16.242 "strip_size_kb": 0, 00:10:16.242 "state": "online", 00:10:16.242 "raid_level": "raid1", 00:10:16.242 "superblock": false, 00:10:16.242 "num_base_bdevs": 3, 00:10:16.242 "num_base_bdevs_discovered": 3, 00:10:16.242 "num_base_bdevs_operational": 3, 00:10:16.242 "base_bdevs_list": [ 00:10:16.242 { 00:10:16.242 "name": "BaseBdev1", 00:10:16.242 "uuid": "ffaf53e2-022f-4529-a0da-aafc64760c9c", 00:10:16.242 "is_configured": true, 00:10:16.242 "data_offset": 0, 00:10:16.242 "data_size": 65536 00:10:16.242 }, 00:10:16.242 { 00:10:16.242 "name": "BaseBdev2", 00:10:16.242 "uuid": "fc8164fc-3b1e-4a0d-bcb3-257ced4f6e9c", 00:10:16.242 "is_configured": true, 00:10:16.242 "data_offset": 0, 00:10:16.242 "data_size": 65536 00:10:16.242 }, 00:10:16.242 { 00:10:16.242 "name": "BaseBdev3", 00:10:16.242 "uuid": "ae268b03-50be-48a4-85b4-3f732353a86b", 00:10:16.242 "is_configured": true, 00:10:16.242 "data_offset": 0, 00:10:16.242 "data_size": 65536 00:10:16.242 } 00:10:16.242 ] 00:10:16.242 }' 00:10:16.242 15:16:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:16.242 15:16:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.501 15:16:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:10:16.501 15:16:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:16.501 15:16:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:16.501 15:16:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:16.501 15:16:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:16.501 15:16:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:16.501 15:16:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:16.501 15:16:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:16.501 15:16:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:16.501 15:16:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.501 [2024-11-19 15:16:06.810797] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:16.501 15:16:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:16.501 15:16:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:16.501 "name": "Existed_Raid", 00:10:16.501 "aliases": [ 00:10:16.501 "35b7e949-0cc9-4592-ab59-bca19d24befc" 00:10:16.501 ], 00:10:16.501 "product_name": "Raid Volume", 00:10:16.501 "block_size": 512, 00:10:16.501 "num_blocks": 65536, 00:10:16.501 "uuid": "35b7e949-0cc9-4592-ab59-bca19d24befc", 00:10:16.501 "assigned_rate_limits": { 00:10:16.501 "rw_ios_per_sec": 0, 00:10:16.501 "rw_mbytes_per_sec": 0, 00:10:16.501 "r_mbytes_per_sec": 0, 00:10:16.501 "w_mbytes_per_sec": 0 00:10:16.501 }, 00:10:16.501 "claimed": false, 00:10:16.501 "zoned": false, 00:10:16.501 "supported_io_types": { 00:10:16.501 "read": true, 00:10:16.501 "write": true, 00:10:16.501 "unmap": false, 00:10:16.501 "flush": false, 00:10:16.501 "reset": true, 00:10:16.501 "nvme_admin": false, 00:10:16.501 "nvme_io": false, 00:10:16.501 "nvme_io_md": false, 00:10:16.501 "write_zeroes": true, 00:10:16.501 "zcopy": false, 00:10:16.501 "get_zone_info": false, 00:10:16.501 "zone_management": false, 00:10:16.501 "zone_append": false, 00:10:16.501 "compare": false, 00:10:16.501 "compare_and_write": false, 00:10:16.501 "abort": false, 00:10:16.501 "seek_hole": false, 00:10:16.501 "seek_data": false, 00:10:16.501 "copy": false, 00:10:16.501 "nvme_iov_md": false 00:10:16.501 }, 00:10:16.501 "memory_domains": [ 00:10:16.501 { 00:10:16.501 "dma_device_id": "system", 00:10:16.501 "dma_device_type": 1 00:10:16.501 }, 00:10:16.501 { 00:10:16.501 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:16.501 "dma_device_type": 2 00:10:16.501 }, 00:10:16.501 { 00:10:16.501 "dma_device_id": "system", 00:10:16.501 "dma_device_type": 1 00:10:16.501 }, 00:10:16.501 { 00:10:16.501 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:16.501 "dma_device_type": 2 00:10:16.501 }, 00:10:16.501 { 00:10:16.501 "dma_device_id": "system", 00:10:16.501 "dma_device_type": 1 00:10:16.501 }, 00:10:16.501 { 00:10:16.501 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:16.501 "dma_device_type": 2 00:10:16.501 } 00:10:16.501 ], 00:10:16.501 "driver_specific": { 00:10:16.501 "raid": { 00:10:16.501 "uuid": "35b7e949-0cc9-4592-ab59-bca19d24befc", 00:10:16.501 "strip_size_kb": 0, 00:10:16.501 "state": "online", 00:10:16.501 "raid_level": "raid1", 00:10:16.501 "superblock": false, 00:10:16.501 "num_base_bdevs": 3, 00:10:16.502 "num_base_bdevs_discovered": 3, 00:10:16.502 "num_base_bdevs_operational": 3, 00:10:16.502 "base_bdevs_list": [ 00:10:16.502 { 00:10:16.502 "name": "BaseBdev1", 00:10:16.502 "uuid": "ffaf53e2-022f-4529-a0da-aafc64760c9c", 00:10:16.502 "is_configured": true, 00:10:16.502 "data_offset": 0, 00:10:16.502 "data_size": 65536 00:10:16.502 }, 00:10:16.502 { 00:10:16.502 "name": "BaseBdev2", 00:10:16.502 "uuid": "fc8164fc-3b1e-4a0d-bcb3-257ced4f6e9c", 00:10:16.502 "is_configured": true, 00:10:16.502 "data_offset": 0, 00:10:16.502 "data_size": 65536 00:10:16.502 }, 00:10:16.502 { 00:10:16.502 "name": "BaseBdev3", 00:10:16.502 "uuid": "ae268b03-50be-48a4-85b4-3f732353a86b", 00:10:16.502 "is_configured": true, 00:10:16.502 "data_offset": 0, 00:10:16.502 "data_size": 65536 00:10:16.502 } 00:10:16.502 ] 00:10:16.502 } 00:10:16.502 } 00:10:16.502 }' 00:10:16.502 15:16:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:16.761 15:16:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:10:16.761 BaseBdev2 00:10:16.761 BaseBdev3' 00:10:16.761 15:16:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:16.761 15:16:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:16.761 15:16:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:16.761 15:16:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:16.761 15:16:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:10:16.761 15:16:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:16.761 15:16:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.761 15:16:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:16.761 15:16:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:16.761 15:16:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:16.761 15:16:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:16.761 15:16:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:16.761 15:16:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:16.761 15:16:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:16.761 15:16:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.761 15:16:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:16.761 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:16.761 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:16.761 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:16.761 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:16.761 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:16.761 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:16.761 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.761 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:16.761 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:16.761 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:16.761 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:16.761 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:16.761 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.761 [2024-11-19 15:16:07.050075] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:16.761 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:16.761 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:10:16.761 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:10:16.761 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:16.761 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:10:16.761 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:10:16.761 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:10:16.761 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:16.761 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:16.761 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:16.761 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:16.761 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:10:16.761 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:16.761 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:16.761 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:16.761 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:16.761 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:16.762 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:16.762 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:16.762 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.762 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.020 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:17.020 "name": "Existed_Raid", 00:10:17.020 "uuid": "35b7e949-0cc9-4592-ab59-bca19d24befc", 00:10:17.020 "strip_size_kb": 0, 00:10:17.020 "state": "online", 00:10:17.020 "raid_level": "raid1", 00:10:17.020 "superblock": false, 00:10:17.020 "num_base_bdevs": 3, 00:10:17.020 "num_base_bdevs_discovered": 2, 00:10:17.020 "num_base_bdevs_operational": 2, 00:10:17.020 "base_bdevs_list": [ 00:10:17.020 { 00:10:17.020 "name": null, 00:10:17.020 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:17.020 "is_configured": false, 00:10:17.020 "data_offset": 0, 00:10:17.020 "data_size": 65536 00:10:17.020 }, 00:10:17.020 { 00:10:17.020 "name": "BaseBdev2", 00:10:17.020 "uuid": "fc8164fc-3b1e-4a0d-bcb3-257ced4f6e9c", 00:10:17.020 "is_configured": true, 00:10:17.020 "data_offset": 0, 00:10:17.020 "data_size": 65536 00:10:17.020 }, 00:10:17.020 { 00:10:17.020 "name": "BaseBdev3", 00:10:17.020 "uuid": "ae268b03-50be-48a4-85b4-3f732353a86b", 00:10:17.020 "is_configured": true, 00:10:17.020 "data_offset": 0, 00:10:17.020 "data_size": 65536 00:10:17.020 } 00:10:17.020 ] 00:10:17.020 }' 00:10:17.020 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:17.020 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.342 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:10:17.342 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:17.342 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:17.342 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.342 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.342 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:17.342 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.342 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:17.342 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:17.342 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:10:17.342 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.342 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.342 [2024-11-19 15:16:07.562264] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:17.342 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.342 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:17.342 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:17.342 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:17.342 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:17.342 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.342 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.342 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.342 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:17.342 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:17.342 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:10:17.342 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.342 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.342 [2024-11-19 15:16:07.639213] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:17.342 [2024-11-19 15:16:07.639334] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:17.601 [2024-11-19 15:16:07.660619] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:17.601 [2024-11-19 15:16:07.660672] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:17.601 [2024-11-19 15:16:07.660690] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:10:17.601 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.601 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:17.601 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:17.601 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:17.601 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.601 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.601 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:10:17.601 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.601 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:10:17.601 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:10:17.601 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:10:17.601 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:10:17.601 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:17.601 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.602 BaseBdev2 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.602 [ 00:10:17.602 { 00:10:17.602 "name": "BaseBdev2", 00:10:17.602 "aliases": [ 00:10:17.602 "fff91556-bac9-4617-8e0a-e69bd004f46b" 00:10:17.602 ], 00:10:17.602 "product_name": "Malloc disk", 00:10:17.602 "block_size": 512, 00:10:17.602 "num_blocks": 65536, 00:10:17.602 "uuid": "fff91556-bac9-4617-8e0a-e69bd004f46b", 00:10:17.602 "assigned_rate_limits": { 00:10:17.602 "rw_ios_per_sec": 0, 00:10:17.602 "rw_mbytes_per_sec": 0, 00:10:17.602 "r_mbytes_per_sec": 0, 00:10:17.602 "w_mbytes_per_sec": 0 00:10:17.602 }, 00:10:17.602 "claimed": false, 00:10:17.602 "zoned": false, 00:10:17.602 "supported_io_types": { 00:10:17.602 "read": true, 00:10:17.602 "write": true, 00:10:17.602 "unmap": true, 00:10:17.602 "flush": true, 00:10:17.602 "reset": true, 00:10:17.602 "nvme_admin": false, 00:10:17.602 "nvme_io": false, 00:10:17.602 "nvme_io_md": false, 00:10:17.602 "write_zeroes": true, 00:10:17.602 "zcopy": true, 00:10:17.602 "get_zone_info": false, 00:10:17.602 "zone_management": false, 00:10:17.602 "zone_append": false, 00:10:17.602 "compare": false, 00:10:17.602 "compare_and_write": false, 00:10:17.602 "abort": true, 00:10:17.602 "seek_hole": false, 00:10:17.602 "seek_data": false, 00:10:17.602 "copy": true, 00:10:17.602 "nvme_iov_md": false 00:10:17.602 }, 00:10:17.602 "memory_domains": [ 00:10:17.602 { 00:10:17.602 "dma_device_id": "system", 00:10:17.602 "dma_device_type": 1 00:10:17.602 }, 00:10:17.602 { 00:10:17.602 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:17.602 "dma_device_type": 2 00:10:17.602 } 00:10:17.602 ], 00:10:17.602 "driver_specific": {} 00:10:17.602 } 00:10:17.602 ] 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.602 BaseBdev3 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.602 [ 00:10:17.602 { 00:10:17.602 "name": "BaseBdev3", 00:10:17.602 "aliases": [ 00:10:17.602 "5f4f2549-9aaf-426c-95c9-cc54620576ca" 00:10:17.602 ], 00:10:17.602 "product_name": "Malloc disk", 00:10:17.602 "block_size": 512, 00:10:17.602 "num_blocks": 65536, 00:10:17.602 "uuid": "5f4f2549-9aaf-426c-95c9-cc54620576ca", 00:10:17.602 "assigned_rate_limits": { 00:10:17.602 "rw_ios_per_sec": 0, 00:10:17.602 "rw_mbytes_per_sec": 0, 00:10:17.602 "r_mbytes_per_sec": 0, 00:10:17.602 "w_mbytes_per_sec": 0 00:10:17.602 }, 00:10:17.602 "claimed": false, 00:10:17.602 "zoned": false, 00:10:17.602 "supported_io_types": { 00:10:17.602 "read": true, 00:10:17.602 "write": true, 00:10:17.602 "unmap": true, 00:10:17.602 "flush": true, 00:10:17.602 "reset": true, 00:10:17.602 "nvme_admin": false, 00:10:17.602 "nvme_io": false, 00:10:17.602 "nvme_io_md": false, 00:10:17.602 "write_zeroes": true, 00:10:17.602 "zcopy": true, 00:10:17.602 "get_zone_info": false, 00:10:17.602 "zone_management": false, 00:10:17.602 "zone_append": false, 00:10:17.602 "compare": false, 00:10:17.602 "compare_and_write": false, 00:10:17.602 "abort": true, 00:10:17.602 "seek_hole": false, 00:10:17.602 "seek_data": false, 00:10:17.602 "copy": true, 00:10:17.602 "nvme_iov_md": false 00:10:17.602 }, 00:10:17.602 "memory_domains": [ 00:10:17.602 { 00:10:17.602 "dma_device_id": "system", 00:10:17.602 "dma_device_type": 1 00:10:17.602 }, 00:10:17.602 { 00:10:17.602 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:17.602 "dma_device_type": 2 00:10:17.602 } 00:10:17.602 ], 00:10:17.602 "driver_specific": {} 00:10:17.602 } 00:10:17.602 ] 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.602 [2024-11-19 15:16:07.836406] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:17.602 [2024-11-19 15:16:07.836470] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:17.602 [2024-11-19 15:16:07.836491] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:17.602 [2024-11-19 15:16:07.838714] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:10:17.602 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:17.603 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:17.603 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:17.603 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:17.603 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:17.603 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:17.603 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:17.603 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:17.603 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:17.603 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:17.603 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:17.603 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.603 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.603 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.603 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:17.603 "name": "Existed_Raid", 00:10:17.603 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:17.603 "strip_size_kb": 0, 00:10:17.603 "state": "configuring", 00:10:17.603 "raid_level": "raid1", 00:10:17.603 "superblock": false, 00:10:17.603 "num_base_bdevs": 3, 00:10:17.603 "num_base_bdevs_discovered": 2, 00:10:17.603 "num_base_bdevs_operational": 3, 00:10:17.603 "base_bdevs_list": [ 00:10:17.603 { 00:10:17.603 "name": "BaseBdev1", 00:10:17.603 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:17.603 "is_configured": false, 00:10:17.603 "data_offset": 0, 00:10:17.603 "data_size": 0 00:10:17.603 }, 00:10:17.603 { 00:10:17.603 "name": "BaseBdev2", 00:10:17.603 "uuid": "fff91556-bac9-4617-8e0a-e69bd004f46b", 00:10:17.603 "is_configured": true, 00:10:17.603 "data_offset": 0, 00:10:17.603 "data_size": 65536 00:10:17.603 }, 00:10:17.603 { 00:10:17.603 "name": "BaseBdev3", 00:10:17.603 "uuid": "5f4f2549-9aaf-426c-95c9-cc54620576ca", 00:10:17.603 "is_configured": true, 00:10:17.603 "data_offset": 0, 00:10:17.603 "data_size": 65536 00:10:17.603 } 00:10:17.603 ] 00:10:17.603 }' 00:10:17.603 15:16:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:17.603 15:16:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.171 15:16:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:10:18.171 15:16:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:18.171 15:16:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.171 [2024-11-19 15:16:08.275720] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:18.171 15:16:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:18.171 15:16:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:10:18.171 15:16:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:18.171 15:16:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:18.171 15:16:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:18.171 15:16:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:18.171 15:16:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:18.171 15:16:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:18.171 15:16:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:18.171 15:16:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:18.171 15:16:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:18.171 15:16:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:18.171 15:16:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:18.171 15:16:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:18.171 15:16:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.171 15:16:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:18.171 15:16:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:18.171 "name": "Existed_Raid", 00:10:18.171 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:18.171 "strip_size_kb": 0, 00:10:18.171 "state": "configuring", 00:10:18.171 "raid_level": "raid1", 00:10:18.171 "superblock": false, 00:10:18.171 "num_base_bdevs": 3, 00:10:18.171 "num_base_bdevs_discovered": 1, 00:10:18.171 "num_base_bdevs_operational": 3, 00:10:18.171 "base_bdevs_list": [ 00:10:18.171 { 00:10:18.171 "name": "BaseBdev1", 00:10:18.171 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:18.171 "is_configured": false, 00:10:18.171 "data_offset": 0, 00:10:18.171 "data_size": 0 00:10:18.171 }, 00:10:18.171 { 00:10:18.171 "name": null, 00:10:18.171 "uuid": "fff91556-bac9-4617-8e0a-e69bd004f46b", 00:10:18.171 "is_configured": false, 00:10:18.171 "data_offset": 0, 00:10:18.171 "data_size": 65536 00:10:18.171 }, 00:10:18.171 { 00:10:18.171 "name": "BaseBdev3", 00:10:18.171 "uuid": "5f4f2549-9aaf-426c-95c9-cc54620576ca", 00:10:18.171 "is_configured": true, 00:10:18.171 "data_offset": 0, 00:10:18.171 "data_size": 65536 00:10:18.171 } 00:10:18.171 ] 00:10:18.171 }' 00:10:18.171 15:16:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:18.171 15:16:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.430 15:16:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:18.430 15:16:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:18.430 15:16:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:18.430 15:16:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.430 15:16:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:18.430 15:16:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:10:18.430 15:16:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:18.430 15:16:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:18.430 15:16:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.430 [2024-11-19 15:16:08.759600] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:18.430 BaseBdev1 00:10:18.430 15:16:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:18.430 15:16:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:10:18.430 15:16:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:10:18.430 15:16:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:18.430 15:16:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:18.430 15:16:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:18.430 15:16:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:18.430 15:16:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:18.430 15:16:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:18.430 15:16:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.689 15:16:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:18.689 15:16:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:18.689 15:16:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:18.689 15:16:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.689 [ 00:10:18.689 { 00:10:18.689 "name": "BaseBdev1", 00:10:18.689 "aliases": [ 00:10:18.689 "d54ff262-6b7b-4999-b6b9-8ae7a16b9af1" 00:10:18.689 ], 00:10:18.689 "product_name": "Malloc disk", 00:10:18.689 "block_size": 512, 00:10:18.689 "num_blocks": 65536, 00:10:18.689 "uuid": "d54ff262-6b7b-4999-b6b9-8ae7a16b9af1", 00:10:18.689 "assigned_rate_limits": { 00:10:18.689 "rw_ios_per_sec": 0, 00:10:18.689 "rw_mbytes_per_sec": 0, 00:10:18.689 "r_mbytes_per_sec": 0, 00:10:18.689 "w_mbytes_per_sec": 0 00:10:18.689 }, 00:10:18.689 "claimed": true, 00:10:18.689 "claim_type": "exclusive_write", 00:10:18.689 "zoned": false, 00:10:18.689 "supported_io_types": { 00:10:18.689 "read": true, 00:10:18.689 "write": true, 00:10:18.689 "unmap": true, 00:10:18.689 "flush": true, 00:10:18.689 "reset": true, 00:10:18.689 "nvme_admin": false, 00:10:18.689 "nvme_io": false, 00:10:18.689 "nvme_io_md": false, 00:10:18.689 "write_zeroes": true, 00:10:18.689 "zcopy": true, 00:10:18.689 "get_zone_info": false, 00:10:18.689 "zone_management": false, 00:10:18.689 "zone_append": false, 00:10:18.689 "compare": false, 00:10:18.689 "compare_and_write": false, 00:10:18.689 "abort": true, 00:10:18.689 "seek_hole": false, 00:10:18.689 "seek_data": false, 00:10:18.689 "copy": true, 00:10:18.689 "nvme_iov_md": false 00:10:18.689 }, 00:10:18.689 "memory_domains": [ 00:10:18.689 { 00:10:18.689 "dma_device_id": "system", 00:10:18.689 "dma_device_type": 1 00:10:18.689 }, 00:10:18.689 { 00:10:18.689 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:18.689 "dma_device_type": 2 00:10:18.689 } 00:10:18.689 ], 00:10:18.689 "driver_specific": {} 00:10:18.689 } 00:10:18.689 ] 00:10:18.689 15:16:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:18.689 15:16:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:18.689 15:16:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:10:18.689 15:16:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:18.689 15:16:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:18.689 15:16:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:18.689 15:16:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:18.689 15:16:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:18.689 15:16:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:18.689 15:16:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:18.689 15:16:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:18.689 15:16:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:18.689 15:16:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:18.689 15:16:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:18.689 15:16:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:18.689 15:16:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.689 15:16:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:18.689 15:16:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:18.689 "name": "Existed_Raid", 00:10:18.689 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:18.689 "strip_size_kb": 0, 00:10:18.689 "state": "configuring", 00:10:18.689 "raid_level": "raid1", 00:10:18.689 "superblock": false, 00:10:18.689 "num_base_bdevs": 3, 00:10:18.689 "num_base_bdevs_discovered": 2, 00:10:18.689 "num_base_bdevs_operational": 3, 00:10:18.689 "base_bdevs_list": [ 00:10:18.689 { 00:10:18.689 "name": "BaseBdev1", 00:10:18.689 "uuid": "d54ff262-6b7b-4999-b6b9-8ae7a16b9af1", 00:10:18.689 "is_configured": true, 00:10:18.689 "data_offset": 0, 00:10:18.689 "data_size": 65536 00:10:18.689 }, 00:10:18.689 { 00:10:18.689 "name": null, 00:10:18.689 "uuid": "fff91556-bac9-4617-8e0a-e69bd004f46b", 00:10:18.689 "is_configured": false, 00:10:18.689 "data_offset": 0, 00:10:18.689 "data_size": 65536 00:10:18.689 }, 00:10:18.689 { 00:10:18.689 "name": "BaseBdev3", 00:10:18.689 "uuid": "5f4f2549-9aaf-426c-95c9-cc54620576ca", 00:10:18.689 "is_configured": true, 00:10:18.689 "data_offset": 0, 00:10:18.689 "data_size": 65536 00:10:18.689 } 00:10:18.689 ] 00:10:18.689 }' 00:10:18.689 15:16:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:18.689 15:16:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.948 15:16:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:18.948 15:16:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:18.948 15:16:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.948 15:16:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:19.207 15:16:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:19.207 15:16:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:10:19.207 15:16:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:10:19.207 15:16:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:19.208 15:16:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.208 [2024-11-19 15:16:09.318710] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:19.208 15:16:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:19.208 15:16:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:10:19.208 15:16:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:19.208 15:16:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:19.208 15:16:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:19.208 15:16:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:19.208 15:16:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:19.208 15:16:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:19.208 15:16:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:19.208 15:16:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:19.208 15:16:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:19.208 15:16:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:19.208 15:16:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:19.208 15:16:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:19.208 15:16:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.208 15:16:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:19.208 15:16:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:19.208 "name": "Existed_Raid", 00:10:19.208 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:19.208 "strip_size_kb": 0, 00:10:19.208 "state": "configuring", 00:10:19.208 "raid_level": "raid1", 00:10:19.208 "superblock": false, 00:10:19.208 "num_base_bdevs": 3, 00:10:19.208 "num_base_bdevs_discovered": 1, 00:10:19.208 "num_base_bdevs_operational": 3, 00:10:19.208 "base_bdevs_list": [ 00:10:19.208 { 00:10:19.208 "name": "BaseBdev1", 00:10:19.208 "uuid": "d54ff262-6b7b-4999-b6b9-8ae7a16b9af1", 00:10:19.208 "is_configured": true, 00:10:19.208 "data_offset": 0, 00:10:19.208 "data_size": 65536 00:10:19.208 }, 00:10:19.208 { 00:10:19.208 "name": null, 00:10:19.208 "uuid": "fff91556-bac9-4617-8e0a-e69bd004f46b", 00:10:19.208 "is_configured": false, 00:10:19.208 "data_offset": 0, 00:10:19.208 "data_size": 65536 00:10:19.208 }, 00:10:19.208 { 00:10:19.208 "name": null, 00:10:19.208 "uuid": "5f4f2549-9aaf-426c-95c9-cc54620576ca", 00:10:19.208 "is_configured": false, 00:10:19.208 "data_offset": 0, 00:10:19.208 "data_size": 65536 00:10:19.208 } 00:10:19.208 ] 00:10:19.208 }' 00:10:19.208 15:16:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:19.208 15:16:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.467 15:16:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:19.467 15:16:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:19.467 15:16:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:19.467 15:16:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.467 15:16:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:19.727 15:16:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:10:19.727 15:16:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:10:19.727 15:16:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:19.727 15:16:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.727 [2024-11-19 15:16:09.813877] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:19.727 15:16:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:19.727 15:16:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:10:19.727 15:16:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:19.727 15:16:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:19.727 15:16:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:19.727 15:16:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:19.727 15:16:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:19.727 15:16:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:19.727 15:16:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:19.727 15:16:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:19.727 15:16:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:19.727 15:16:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:19.727 15:16:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:19.727 15:16:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:19.727 15:16:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.727 15:16:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:19.727 15:16:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:19.727 "name": "Existed_Raid", 00:10:19.727 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:19.727 "strip_size_kb": 0, 00:10:19.727 "state": "configuring", 00:10:19.727 "raid_level": "raid1", 00:10:19.727 "superblock": false, 00:10:19.727 "num_base_bdevs": 3, 00:10:19.727 "num_base_bdevs_discovered": 2, 00:10:19.727 "num_base_bdevs_operational": 3, 00:10:19.727 "base_bdevs_list": [ 00:10:19.727 { 00:10:19.727 "name": "BaseBdev1", 00:10:19.727 "uuid": "d54ff262-6b7b-4999-b6b9-8ae7a16b9af1", 00:10:19.727 "is_configured": true, 00:10:19.727 "data_offset": 0, 00:10:19.727 "data_size": 65536 00:10:19.727 }, 00:10:19.727 { 00:10:19.727 "name": null, 00:10:19.727 "uuid": "fff91556-bac9-4617-8e0a-e69bd004f46b", 00:10:19.727 "is_configured": false, 00:10:19.727 "data_offset": 0, 00:10:19.727 "data_size": 65536 00:10:19.727 }, 00:10:19.727 { 00:10:19.727 "name": "BaseBdev3", 00:10:19.727 "uuid": "5f4f2549-9aaf-426c-95c9-cc54620576ca", 00:10:19.727 "is_configured": true, 00:10:19.727 "data_offset": 0, 00:10:19.727 "data_size": 65536 00:10:19.727 } 00:10:19.727 ] 00:10:19.727 }' 00:10:19.727 15:16:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:19.727 15:16:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.987 15:16:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:19.987 15:16:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:19.987 15:16:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:19.987 15:16:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.987 15:16:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:19.987 15:16:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:10:19.987 15:16:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:19.987 15:16:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:19.987 15:16:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.987 [2024-11-19 15:16:10.225209] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:19.987 15:16:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:19.987 15:16:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:10:19.987 15:16:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:19.987 15:16:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:19.987 15:16:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:19.987 15:16:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:19.987 15:16:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:19.987 15:16:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:19.987 15:16:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:19.987 15:16:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:19.987 15:16:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:19.987 15:16:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:19.987 15:16:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:19.987 15:16:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.987 15:16:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:19.987 15:16:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:19.987 15:16:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:19.987 "name": "Existed_Raid", 00:10:19.987 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:19.987 "strip_size_kb": 0, 00:10:19.987 "state": "configuring", 00:10:19.987 "raid_level": "raid1", 00:10:19.987 "superblock": false, 00:10:19.987 "num_base_bdevs": 3, 00:10:19.987 "num_base_bdevs_discovered": 1, 00:10:19.987 "num_base_bdevs_operational": 3, 00:10:19.987 "base_bdevs_list": [ 00:10:19.987 { 00:10:19.987 "name": null, 00:10:19.987 "uuid": "d54ff262-6b7b-4999-b6b9-8ae7a16b9af1", 00:10:19.987 "is_configured": false, 00:10:19.987 "data_offset": 0, 00:10:19.987 "data_size": 65536 00:10:19.987 }, 00:10:19.987 { 00:10:19.987 "name": null, 00:10:19.987 "uuid": "fff91556-bac9-4617-8e0a-e69bd004f46b", 00:10:19.987 "is_configured": false, 00:10:19.987 "data_offset": 0, 00:10:19.987 "data_size": 65536 00:10:19.987 }, 00:10:19.987 { 00:10:19.987 "name": "BaseBdev3", 00:10:19.987 "uuid": "5f4f2549-9aaf-426c-95c9-cc54620576ca", 00:10:19.987 "is_configured": true, 00:10:19.987 "data_offset": 0, 00:10:19.987 "data_size": 65536 00:10:19.987 } 00:10:19.987 ] 00:10:19.987 }' 00:10:19.987 15:16:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:19.987 15:16:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.556 15:16:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:20.556 15:16:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:20.556 15:16:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:20.556 15:16:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.556 15:16:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:20.556 15:16:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:10:20.556 15:16:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:10:20.556 15:16:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:20.556 15:16:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.556 [2024-11-19 15:16:10.692747] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:20.556 15:16:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:20.556 15:16:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:10:20.556 15:16:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:20.556 15:16:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:20.556 15:16:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:20.556 15:16:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:20.556 15:16:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:20.556 15:16:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:20.556 15:16:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:20.556 15:16:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:20.556 15:16:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:20.556 15:16:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:20.556 15:16:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:20.556 15:16:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:20.556 15:16:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.556 15:16:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:20.556 15:16:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:20.556 "name": "Existed_Raid", 00:10:20.556 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:20.556 "strip_size_kb": 0, 00:10:20.556 "state": "configuring", 00:10:20.556 "raid_level": "raid1", 00:10:20.556 "superblock": false, 00:10:20.556 "num_base_bdevs": 3, 00:10:20.556 "num_base_bdevs_discovered": 2, 00:10:20.556 "num_base_bdevs_operational": 3, 00:10:20.556 "base_bdevs_list": [ 00:10:20.556 { 00:10:20.556 "name": null, 00:10:20.556 "uuid": "d54ff262-6b7b-4999-b6b9-8ae7a16b9af1", 00:10:20.556 "is_configured": false, 00:10:20.556 "data_offset": 0, 00:10:20.556 "data_size": 65536 00:10:20.556 }, 00:10:20.556 { 00:10:20.556 "name": "BaseBdev2", 00:10:20.556 "uuid": "fff91556-bac9-4617-8e0a-e69bd004f46b", 00:10:20.556 "is_configured": true, 00:10:20.556 "data_offset": 0, 00:10:20.556 "data_size": 65536 00:10:20.556 }, 00:10:20.556 { 00:10:20.556 "name": "BaseBdev3", 00:10:20.556 "uuid": "5f4f2549-9aaf-426c-95c9-cc54620576ca", 00:10:20.556 "is_configured": true, 00:10:20.556 "data_offset": 0, 00:10:20.556 "data_size": 65536 00:10:20.556 } 00:10:20.556 ] 00:10:20.556 }' 00:10:20.556 15:16:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:20.556 15:16:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.816 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:20.816 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:20.816 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:20.816 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.816 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.076 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:10:21.076 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:21.076 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.076 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.076 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:10:21.076 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.076 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u d54ff262-6b7b-4999-b6b9-8ae7a16b9af1 00:10:21.076 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.076 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.076 [2024-11-19 15:16:11.244819] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:10:21.076 [2024-11-19 15:16:11.244881] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:10:21.076 [2024-11-19 15:16:11.244890] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:10:21.076 [2024-11-19 15:16:11.245238] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:10:21.076 [2024-11-19 15:16:11.245389] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:10:21.076 [2024-11-19 15:16:11.245410] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:10:21.076 [2024-11-19 15:16:11.245617] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:21.076 NewBaseBdev 00:10:21.076 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.076 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:10:21.076 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:10:21.076 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:21.076 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:21.076 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:21.076 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:21.076 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:21.076 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.076 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.076 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.076 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:10:21.076 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.076 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.076 [ 00:10:21.076 { 00:10:21.076 "name": "NewBaseBdev", 00:10:21.076 "aliases": [ 00:10:21.076 "d54ff262-6b7b-4999-b6b9-8ae7a16b9af1" 00:10:21.076 ], 00:10:21.076 "product_name": "Malloc disk", 00:10:21.076 "block_size": 512, 00:10:21.076 "num_blocks": 65536, 00:10:21.076 "uuid": "d54ff262-6b7b-4999-b6b9-8ae7a16b9af1", 00:10:21.076 "assigned_rate_limits": { 00:10:21.076 "rw_ios_per_sec": 0, 00:10:21.076 "rw_mbytes_per_sec": 0, 00:10:21.076 "r_mbytes_per_sec": 0, 00:10:21.076 "w_mbytes_per_sec": 0 00:10:21.076 }, 00:10:21.076 "claimed": true, 00:10:21.076 "claim_type": "exclusive_write", 00:10:21.076 "zoned": false, 00:10:21.076 "supported_io_types": { 00:10:21.076 "read": true, 00:10:21.076 "write": true, 00:10:21.076 "unmap": true, 00:10:21.076 "flush": true, 00:10:21.076 "reset": true, 00:10:21.076 "nvme_admin": false, 00:10:21.076 "nvme_io": false, 00:10:21.076 "nvme_io_md": false, 00:10:21.076 "write_zeroes": true, 00:10:21.076 "zcopy": true, 00:10:21.076 "get_zone_info": false, 00:10:21.076 "zone_management": false, 00:10:21.076 "zone_append": false, 00:10:21.076 "compare": false, 00:10:21.076 "compare_and_write": false, 00:10:21.076 "abort": true, 00:10:21.076 "seek_hole": false, 00:10:21.076 "seek_data": false, 00:10:21.076 "copy": true, 00:10:21.076 "nvme_iov_md": false 00:10:21.076 }, 00:10:21.076 "memory_domains": [ 00:10:21.076 { 00:10:21.076 "dma_device_id": "system", 00:10:21.076 "dma_device_type": 1 00:10:21.076 }, 00:10:21.076 { 00:10:21.076 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:21.076 "dma_device_type": 2 00:10:21.076 } 00:10:21.076 ], 00:10:21.076 "driver_specific": {} 00:10:21.076 } 00:10:21.076 ] 00:10:21.076 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.076 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:21.076 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:10:21.076 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:21.076 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:21.076 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:21.076 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:21.077 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:21.077 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:21.077 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:21.077 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:21.077 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:21.077 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:21.077 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.077 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:21.077 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.077 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.077 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:21.077 "name": "Existed_Raid", 00:10:21.077 "uuid": "cc88efc6-b16c-43d6-b0c3-ddc2de0e97d6", 00:10:21.077 "strip_size_kb": 0, 00:10:21.077 "state": "online", 00:10:21.077 "raid_level": "raid1", 00:10:21.077 "superblock": false, 00:10:21.077 "num_base_bdevs": 3, 00:10:21.077 "num_base_bdevs_discovered": 3, 00:10:21.077 "num_base_bdevs_operational": 3, 00:10:21.077 "base_bdevs_list": [ 00:10:21.077 { 00:10:21.077 "name": "NewBaseBdev", 00:10:21.077 "uuid": "d54ff262-6b7b-4999-b6b9-8ae7a16b9af1", 00:10:21.077 "is_configured": true, 00:10:21.077 "data_offset": 0, 00:10:21.077 "data_size": 65536 00:10:21.077 }, 00:10:21.077 { 00:10:21.077 "name": "BaseBdev2", 00:10:21.077 "uuid": "fff91556-bac9-4617-8e0a-e69bd004f46b", 00:10:21.077 "is_configured": true, 00:10:21.077 "data_offset": 0, 00:10:21.077 "data_size": 65536 00:10:21.077 }, 00:10:21.077 { 00:10:21.077 "name": "BaseBdev3", 00:10:21.077 "uuid": "5f4f2549-9aaf-426c-95c9-cc54620576ca", 00:10:21.077 "is_configured": true, 00:10:21.077 "data_offset": 0, 00:10:21.077 "data_size": 65536 00:10:21.077 } 00:10:21.077 ] 00:10:21.077 }' 00:10:21.077 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:21.077 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.647 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:10:21.647 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:21.647 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:21.647 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:21.647 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:21.647 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:21.647 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:21.647 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:21.647 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.647 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.647 [2024-11-19 15:16:11.704491] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:21.647 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.647 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:21.647 "name": "Existed_Raid", 00:10:21.647 "aliases": [ 00:10:21.647 "cc88efc6-b16c-43d6-b0c3-ddc2de0e97d6" 00:10:21.647 ], 00:10:21.647 "product_name": "Raid Volume", 00:10:21.647 "block_size": 512, 00:10:21.647 "num_blocks": 65536, 00:10:21.647 "uuid": "cc88efc6-b16c-43d6-b0c3-ddc2de0e97d6", 00:10:21.647 "assigned_rate_limits": { 00:10:21.647 "rw_ios_per_sec": 0, 00:10:21.647 "rw_mbytes_per_sec": 0, 00:10:21.647 "r_mbytes_per_sec": 0, 00:10:21.647 "w_mbytes_per_sec": 0 00:10:21.647 }, 00:10:21.647 "claimed": false, 00:10:21.647 "zoned": false, 00:10:21.647 "supported_io_types": { 00:10:21.647 "read": true, 00:10:21.647 "write": true, 00:10:21.647 "unmap": false, 00:10:21.647 "flush": false, 00:10:21.647 "reset": true, 00:10:21.647 "nvme_admin": false, 00:10:21.647 "nvme_io": false, 00:10:21.647 "nvme_io_md": false, 00:10:21.647 "write_zeroes": true, 00:10:21.647 "zcopy": false, 00:10:21.647 "get_zone_info": false, 00:10:21.647 "zone_management": false, 00:10:21.647 "zone_append": false, 00:10:21.647 "compare": false, 00:10:21.647 "compare_and_write": false, 00:10:21.647 "abort": false, 00:10:21.647 "seek_hole": false, 00:10:21.647 "seek_data": false, 00:10:21.647 "copy": false, 00:10:21.647 "nvme_iov_md": false 00:10:21.647 }, 00:10:21.647 "memory_domains": [ 00:10:21.647 { 00:10:21.647 "dma_device_id": "system", 00:10:21.647 "dma_device_type": 1 00:10:21.647 }, 00:10:21.647 { 00:10:21.647 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:21.647 "dma_device_type": 2 00:10:21.647 }, 00:10:21.647 { 00:10:21.647 "dma_device_id": "system", 00:10:21.647 "dma_device_type": 1 00:10:21.647 }, 00:10:21.647 { 00:10:21.647 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:21.647 "dma_device_type": 2 00:10:21.647 }, 00:10:21.647 { 00:10:21.647 "dma_device_id": "system", 00:10:21.647 "dma_device_type": 1 00:10:21.647 }, 00:10:21.647 { 00:10:21.647 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:21.647 "dma_device_type": 2 00:10:21.647 } 00:10:21.647 ], 00:10:21.647 "driver_specific": { 00:10:21.647 "raid": { 00:10:21.647 "uuid": "cc88efc6-b16c-43d6-b0c3-ddc2de0e97d6", 00:10:21.647 "strip_size_kb": 0, 00:10:21.647 "state": "online", 00:10:21.647 "raid_level": "raid1", 00:10:21.647 "superblock": false, 00:10:21.647 "num_base_bdevs": 3, 00:10:21.647 "num_base_bdevs_discovered": 3, 00:10:21.647 "num_base_bdevs_operational": 3, 00:10:21.647 "base_bdevs_list": [ 00:10:21.647 { 00:10:21.647 "name": "NewBaseBdev", 00:10:21.647 "uuid": "d54ff262-6b7b-4999-b6b9-8ae7a16b9af1", 00:10:21.647 "is_configured": true, 00:10:21.647 "data_offset": 0, 00:10:21.647 "data_size": 65536 00:10:21.647 }, 00:10:21.647 { 00:10:21.647 "name": "BaseBdev2", 00:10:21.647 "uuid": "fff91556-bac9-4617-8e0a-e69bd004f46b", 00:10:21.647 "is_configured": true, 00:10:21.647 "data_offset": 0, 00:10:21.647 "data_size": 65536 00:10:21.647 }, 00:10:21.647 { 00:10:21.647 "name": "BaseBdev3", 00:10:21.647 "uuid": "5f4f2549-9aaf-426c-95c9-cc54620576ca", 00:10:21.647 "is_configured": true, 00:10:21.647 "data_offset": 0, 00:10:21.647 "data_size": 65536 00:10:21.647 } 00:10:21.647 ] 00:10:21.647 } 00:10:21.647 } 00:10:21.647 }' 00:10:21.647 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:21.647 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:10:21.647 BaseBdev2 00:10:21.647 BaseBdev3' 00:10:21.647 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:21.647 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:21.647 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:21.647 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:10:21.647 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:21.647 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.647 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.647 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.647 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:21.647 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:21.647 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:21.647 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:21.647 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.647 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.647 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:21.647 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.647 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:21.647 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:21.647 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:21.647 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:21.647 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:21.647 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.647 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.647 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.647 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:21.647 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:21.647 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:21.647 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.647 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.907 [2024-11-19 15:16:11.987626] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:21.907 [2024-11-19 15:16:11.987672] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:21.907 [2024-11-19 15:16:11.987770] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:21.907 [2024-11-19 15:16:11.988100] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:21.908 [2024-11-19 15:16:11.988122] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:10:21.908 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.908 15:16:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 78454 00:10:21.908 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 78454 ']' 00:10:21.908 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 78454 00:10:21.908 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:10:21.908 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:21.908 15:16:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 78454 00:10:21.908 15:16:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:21.908 15:16:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:21.908 killing process with pid 78454 00:10:21.908 15:16:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 78454' 00:10:21.908 15:16:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 78454 00:10:21.908 [2024-11-19 15:16:12.031337] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:21.908 15:16:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 78454 00:10:21.908 [2024-11-19 15:16:12.091603] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:22.167 15:16:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:10:22.167 00:10:22.167 real 0m8.920s 00:10:22.167 user 0m15.020s 00:10:22.167 sys 0m1.826s 00:10:22.167 15:16:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:22.167 15:16:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.167 ************************************ 00:10:22.167 END TEST raid_state_function_test 00:10:22.167 ************************************ 00:10:22.167 15:16:12 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid1 3 true 00:10:22.167 15:16:12 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:10:22.167 15:16:12 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:22.167 15:16:12 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:22.167 ************************************ 00:10:22.167 START TEST raid_state_function_test_sb 00:10:22.167 ************************************ 00:10:22.167 15:16:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 3 true 00:10:22.167 15:16:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:10:22.167 15:16:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:10:22.167 15:16:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:10:22.167 15:16:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:10:22.167 15:16:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:10:22.167 15:16:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:22.167 15:16:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:10:22.167 15:16:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:22.167 15:16:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:22.167 15:16:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:10:22.167 15:16:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:22.167 15:16:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:22.167 15:16:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:10:22.167 15:16:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:22.167 15:16:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:22.167 15:16:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:10:22.167 15:16:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:10:22.167 15:16:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:10:22.167 15:16:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:10:22.167 15:16:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:10:22.167 15:16:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:10:22.167 15:16:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:10:22.167 15:16:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:10:22.167 15:16:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:10:22.167 15:16:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:10:22.167 15:16:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=79053 00:10:22.167 15:16:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:10:22.167 15:16:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 79053' 00:10:22.167 Process raid pid: 79053 00:10:22.167 15:16:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 79053 00:10:22.167 15:16:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 79053 ']' 00:10:22.167 15:16:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:22.167 15:16:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:22.167 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:22.167 15:16:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:22.167 15:16:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:22.167 15:16:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:22.427 [2024-11-19 15:16:12.576047] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:10:22.427 [2024-11-19 15:16:12.576174] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:22.427 [2024-11-19 15:16:12.730777] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:22.686 [2024-11-19 15:16:12.770736] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:22.686 [2024-11-19 15:16:12.846995] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:22.686 [2024-11-19 15:16:12.847056] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:23.255 15:16:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:23.255 15:16:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:10:23.255 15:16:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:10:23.255 15:16:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:23.255 15:16:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:23.255 [2024-11-19 15:16:13.430350] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:23.255 [2024-11-19 15:16:13.430429] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:23.255 [2024-11-19 15:16:13.430440] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:23.255 [2024-11-19 15:16:13.430452] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:23.255 [2024-11-19 15:16:13.430458] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:23.255 [2024-11-19 15:16:13.430471] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:23.255 15:16:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:23.255 15:16:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:10:23.255 15:16:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:23.255 15:16:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:23.255 15:16:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:23.255 15:16:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:23.255 15:16:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:23.255 15:16:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:23.255 15:16:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:23.255 15:16:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:23.255 15:16:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:23.255 15:16:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:23.255 15:16:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:23.255 15:16:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:23.255 15:16:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:23.255 15:16:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:23.255 15:16:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:23.255 "name": "Existed_Raid", 00:10:23.255 "uuid": "69b69690-2444-47bb-943c-a58b666ef7b0", 00:10:23.255 "strip_size_kb": 0, 00:10:23.255 "state": "configuring", 00:10:23.255 "raid_level": "raid1", 00:10:23.255 "superblock": true, 00:10:23.255 "num_base_bdevs": 3, 00:10:23.255 "num_base_bdevs_discovered": 0, 00:10:23.255 "num_base_bdevs_operational": 3, 00:10:23.255 "base_bdevs_list": [ 00:10:23.255 { 00:10:23.255 "name": "BaseBdev1", 00:10:23.255 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:23.255 "is_configured": false, 00:10:23.255 "data_offset": 0, 00:10:23.255 "data_size": 0 00:10:23.255 }, 00:10:23.255 { 00:10:23.255 "name": "BaseBdev2", 00:10:23.255 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:23.255 "is_configured": false, 00:10:23.255 "data_offset": 0, 00:10:23.255 "data_size": 0 00:10:23.255 }, 00:10:23.255 { 00:10:23.255 "name": "BaseBdev3", 00:10:23.255 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:23.255 "is_configured": false, 00:10:23.255 "data_offset": 0, 00:10:23.255 "data_size": 0 00:10:23.255 } 00:10:23.255 ] 00:10:23.255 }' 00:10:23.255 15:16:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:23.255 15:16:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:23.824 15:16:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:23.824 15:16:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:23.824 15:16:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:23.824 [2024-11-19 15:16:13.893551] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:23.824 [2024-11-19 15:16:13.893617] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:10:23.824 15:16:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:23.824 15:16:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:10:23.824 15:16:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:23.824 15:16:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:23.824 [2024-11-19 15:16:13.905511] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:23.824 [2024-11-19 15:16:13.905555] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:23.824 [2024-11-19 15:16:13.905564] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:23.824 [2024-11-19 15:16:13.905575] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:23.824 [2024-11-19 15:16:13.905581] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:23.825 [2024-11-19 15:16:13.905590] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:23.825 15:16:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:23.825 15:16:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:23.825 15:16:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:23.825 15:16:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:23.825 [2024-11-19 15:16:13.932396] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:23.825 BaseBdev1 00:10:23.825 15:16:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:23.825 15:16:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:10:23.825 15:16:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:10:23.825 15:16:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:23.825 15:16:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:23.825 15:16:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:23.825 15:16:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:23.825 15:16:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:23.825 15:16:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:23.825 15:16:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:23.825 15:16:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:23.825 15:16:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:23.825 15:16:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:23.825 15:16:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:23.825 [ 00:10:23.825 { 00:10:23.825 "name": "BaseBdev1", 00:10:23.825 "aliases": [ 00:10:23.825 "aee8f599-3136-49ab-b1fb-1d770643e893" 00:10:23.825 ], 00:10:23.825 "product_name": "Malloc disk", 00:10:23.825 "block_size": 512, 00:10:23.825 "num_blocks": 65536, 00:10:23.825 "uuid": "aee8f599-3136-49ab-b1fb-1d770643e893", 00:10:23.825 "assigned_rate_limits": { 00:10:23.825 "rw_ios_per_sec": 0, 00:10:23.825 "rw_mbytes_per_sec": 0, 00:10:23.825 "r_mbytes_per_sec": 0, 00:10:23.825 "w_mbytes_per_sec": 0 00:10:23.825 }, 00:10:23.825 "claimed": true, 00:10:23.825 "claim_type": "exclusive_write", 00:10:23.825 "zoned": false, 00:10:23.825 "supported_io_types": { 00:10:23.825 "read": true, 00:10:23.825 "write": true, 00:10:23.825 "unmap": true, 00:10:23.825 "flush": true, 00:10:23.825 "reset": true, 00:10:23.825 "nvme_admin": false, 00:10:23.825 "nvme_io": false, 00:10:23.825 "nvme_io_md": false, 00:10:23.825 "write_zeroes": true, 00:10:23.825 "zcopy": true, 00:10:23.825 "get_zone_info": false, 00:10:23.825 "zone_management": false, 00:10:23.825 "zone_append": false, 00:10:23.825 "compare": false, 00:10:23.825 "compare_and_write": false, 00:10:23.825 "abort": true, 00:10:23.825 "seek_hole": false, 00:10:23.825 "seek_data": false, 00:10:23.825 "copy": true, 00:10:23.825 "nvme_iov_md": false 00:10:23.825 }, 00:10:23.825 "memory_domains": [ 00:10:23.825 { 00:10:23.825 "dma_device_id": "system", 00:10:23.825 "dma_device_type": 1 00:10:23.825 }, 00:10:23.825 { 00:10:23.825 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:23.825 "dma_device_type": 2 00:10:23.825 } 00:10:23.825 ], 00:10:23.825 "driver_specific": {} 00:10:23.825 } 00:10:23.825 ] 00:10:23.825 15:16:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:23.825 15:16:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:23.825 15:16:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:10:23.825 15:16:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:23.825 15:16:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:23.825 15:16:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:23.825 15:16:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:23.825 15:16:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:23.825 15:16:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:23.825 15:16:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:23.825 15:16:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:23.825 15:16:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:23.825 15:16:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:23.825 15:16:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:23.825 15:16:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:23.825 15:16:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:23.825 15:16:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:23.825 15:16:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:23.825 "name": "Existed_Raid", 00:10:23.825 "uuid": "e96f603b-9298-454d-b51e-0f43fcd4a60f", 00:10:23.825 "strip_size_kb": 0, 00:10:23.825 "state": "configuring", 00:10:23.825 "raid_level": "raid1", 00:10:23.825 "superblock": true, 00:10:23.825 "num_base_bdevs": 3, 00:10:23.825 "num_base_bdevs_discovered": 1, 00:10:23.825 "num_base_bdevs_operational": 3, 00:10:23.825 "base_bdevs_list": [ 00:10:23.825 { 00:10:23.825 "name": "BaseBdev1", 00:10:23.825 "uuid": "aee8f599-3136-49ab-b1fb-1d770643e893", 00:10:23.825 "is_configured": true, 00:10:23.825 "data_offset": 2048, 00:10:23.825 "data_size": 63488 00:10:23.825 }, 00:10:23.825 { 00:10:23.825 "name": "BaseBdev2", 00:10:23.825 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:23.825 "is_configured": false, 00:10:23.825 "data_offset": 0, 00:10:23.825 "data_size": 0 00:10:23.825 }, 00:10:23.825 { 00:10:23.825 "name": "BaseBdev3", 00:10:23.825 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:23.825 "is_configured": false, 00:10:23.825 "data_offset": 0, 00:10:23.825 "data_size": 0 00:10:23.825 } 00:10:23.825 ] 00:10:23.825 }' 00:10:23.825 15:16:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:23.825 15:16:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:24.085 15:16:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:24.085 15:16:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:24.085 15:16:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:24.085 [2024-11-19 15:16:14.383803] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:24.085 [2024-11-19 15:16:14.383892] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:10:24.085 15:16:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:24.085 15:16:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:10:24.086 15:16:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:24.086 15:16:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:24.086 [2024-11-19 15:16:14.391782] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:24.086 [2024-11-19 15:16:14.394000] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:24.086 [2024-11-19 15:16:14.394042] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:24.086 [2024-11-19 15:16:14.394051] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:24.086 [2024-11-19 15:16:14.394062] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:24.086 15:16:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:24.086 15:16:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:10:24.086 15:16:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:24.086 15:16:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:10:24.086 15:16:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:24.086 15:16:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:24.086 15:16:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:24.086 15:16:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:24.086 15:16:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:24.086 15:16:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:24.086 15:16:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:24.086 15:16:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:24.086 15:16:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:24.086 15:16:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:24.086 15:16:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:24.086 15:16:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:24.086 15:16:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:24.346 15:16:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:24.346 15:16:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:24.346 "name": "Existed_Raid", 00:10:24.346 "uuid": "01327759-8873-4538-8fea-36c3305ff439", 00:10:24.346 "strip_size_kb": 0, 00:10:24.346 "state": "configuring", 00:10:24.346 "raid_level": "raid1", 00:10:24.346 "superblock": true, 00:10:24.346 "num_base_bdevs": 3, 00:10:24.346 "num_base_bdevs_discovered": 1, 00:10:24.346 "num_base_bdevs_operational": 3, 00:10:24.346 "base_bdevs_list": [ 00:10:24.346 { 00:10:24.346 "name": "BaseBdev1", 00:10:24.346 "uuid": "aee8f599-3136-49ab-b1fb-1d770643e893", 00:10:24.346 "is_configured": true, 00:10:24.346 "data_offset": 2048, 00:10:24.346 "data_size": 63488 00:10:24.346 }, 00:10:24.346 { 00:10:24.346 "name": "BaseBdev2", 00:10:24.346 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:24.346 "is_configured": false, 00:10:24.346 "data_offset": 0, 00:10:24.346 "data_size": 0 00:10:24.346 }, 00:10:24.346 { 00:10:24.346 "name": "BaseBdev3", 00:10:24.346 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:24.346 "is_configured": false, 00:10:24.346 "data_offset": 0, 00:10:24.346 "data_size": 0 00:10:24.346 } 00:10:24.346 ] 00:10:24.346 }' 00:10:24.346 15:16:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:24.346 15:16:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:24.606 15:16:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:24.606 15:16:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:24.606 15:16:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:24.606 [2024-11-19 15:16:14.875908] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:24.606 BaseBdev2 00:10:24.606 15:16:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:24.606 15:16:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:10:24.606 15:16:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:10:24.606 15:16:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:24.606 15:16:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:24.606 15:16:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:24.606 15:16:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:24.606 15:16:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:24.606 15:16:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:24.606 15:16:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:24.606 15:16:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:24.606 15:16:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:24.606 15:16:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:24.606 15:16:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:24.606 [ 00:10:24.606 { 00:10:24.606 "name": "BaseBdev2", 00:10:24.606 "aliases": [ 00:10:24.606 "168466db-c5dc-448f-a7ad-142fe4b607ca" 00:10:24.606 ], 00:10:24.606 "product_name": "Malloc disk", 00:10:24.606 "block_size": 512, 00:10:24.606 "num_blocks": 65536, 00:10:24.607 "uuid": "168466db-c5dc-448f-a7ad-142fe4b607ca", 00:10:24.607 "assigned_rate_limits": { 00:10:24.607 "rw_ios_per_sec": 0, 00:10:24.607 "rw_mbytes_per_sec": 0, 00:10:24.607 "r_mbytes_per_sec": 0, 00:10:24.607 "w_mbytes_per_sec": 0 00:10:24.607 }, 00:10:24.607 "claimed": true, 00:10:24.607 "claim_type": "exclusive_write", 00:10:24.607 "zoned": false, 00:10:24.607 "supported_io_types": { 00:10:24.607 "read": true, 00:10:24.607 "write": true, 00:10:24.607 "unmap": true, 00:10:24.607 "flush": true, 00:10:24.607 "reset": true, 00:10:24.607 "nvme_admin": false, 00:10:24.607 "nvme_io": false, 00:10:24.607 "nvme_io_md": false, 00:10:24.607 "write_zeroes": true, 00:10:24.607 "zcopy": true, 00:10:24.607 "get_zone_info": false, 00:10:24.607 "zone_management": false, 00:10:24.607 "zone_append": false, 00:10:24.607 "compare": false, 00:10:24.607 "compare_and_write": false, 00:10:24.607 "abort": true, 00:10:24.607 "seek_hole": false, 00:10:24.607 "seek_data": false, 00:10:24.607 "copy": true, 00:10:24.607 "nvme_iov_md": false 00:10:24.607 }, 00:10:24.607 "memory_domains": [ 00:10:24.607 { 00:10:24.607 "dma_device_id": "system", 00:10:24.607 "dma_device_type": 1 00:10:24.607 }, 00:10:24.607 { 00:10:24.607 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:24.607 "dma_device_type": 2 00:10:24.607 } 00:10:24.607 ], 00:10:24.607 "driver_specific": {} 00:10:24.607 } 00:10:24.607 ] 00:10:24.607 15:16:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:24.607 15:16:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:24.607 15:16:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:24.607 15:16:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:24.607 15:16:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:10:24.607 15:16:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:24.607 15:16:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:24.607 15:16:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:24.607 15:16:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:24.607 15:16:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:24.607 15:16:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:24.607 15:16:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:24.607 15:16:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:24.607 15:16:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:24.607 15:16:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:24.607 15:16:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:24.607 15:16:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:24.607 15:16:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:24.607 15:16:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:24.867 15:16:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:24.867 "name": "Existed_Raid", 00:10:24.867 "uuid": "01327759-8873-4538-8fea-36c3305ff439", 00:10:24.867 "strip_size_kb": 0, 00:10:24.867 "state": "configuring", 00:10:24.867 "raid_level": "raid1", 00:10:24.867 "superblock": true, 00:10:24.867 "num_base_bdevs": 3, 00:10:24.867 "num_base_bdevs_discovered": 2, 00:10:24.867 "num_base_bdevs_operational": 3, 00:10:24.867 "base_bdevs_list": [ 00:10:24.867 { 00:10:24.867 "name": "BaseBdev1", 00:10:24.867 "uuid": "aee8f599-3136-49ab-b1fb-1d770643e893", 00:10:24.867 "is_configured": true, 00:10:24.867 "data_offset": 2048, 00:10:24.867 "data_size": 63488 00:10:24.867 }, 00:10:24.867 { 00:10:24.867 "name": "BaseBdev2", 00:10:24.867 "uuid": "168466db-c5dc-448f-a7ad-142fe4b607ca", 00:10:24.867 "is_configured": true, 00:10:24.867 "data_offset": 2048, 00:10:24.867 "data_size": 63488 00:10:24.867 }, 00:10:24.867 { 00:10:24.867 "name": "BaseBdev3", 00:10:24.867 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:24.867 "is_configured": false, 00:10:24.867 "data_offset": 0, 00:10:24.867 "data_size": 0 00:10:24.867 } 00:10:24.867 ] 00:10:24.867 }' 00:10:24.867 15:16:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:24.867 15:16:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:25.130 15:16:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:25.130 15:16:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.130 15:16:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:25.130 [2024-11-19 15:16:15.377694] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:25.130 [2024-11-19 15:16:15.378070] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:10:25.130 [2024-11-19 15:16:15.378116] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:25.130 BaseBdev3 00:10:25.130 [2024-11-19 15:16:15.378693] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:10:25.130 [2024-11-19 15:16:15.379018] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:10:25.130 [2024-11-19 15:16:15.379050] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:10:25.130 [2024-11-19 15:16:15.379326] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:25.130 15:16:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.130 15:16:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:10:25.130 15:16:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:10:25.130 15:16:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:25.130 15:16:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:25.130 15:16:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:25.130 15:16:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:25.130 15:16:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:25.130 15:16:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.130 15:16:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:25.130 15:16:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.130 15:16:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:25.130 15:16:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.130 15:16:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:25.130 [ 00:10:25.130 { 00:10:25.130 "name": "BaseBdev3", 00:10:25.130 "aliases": [ 00:10:25.130 "005bfa81-fe96-4307-b30c-bfa6d8ac685c" 00:10:25.130 ], 00:10:25.130 "product_name": "Malloc disk", 00:10:25.130 "block_size": 512, 00:10:25.130 "num_blocks": 65536, 00:10:25.130 "uuid": "005bfa81-fe96-4307-b30c-bfa6d8ac685c", 00:10:25.130 "assigned_rate_limits": { 00:10:25.130 "rw_ios_per_sec": 0, 00:10:25.130 "rw_mbytes_per_sec": 0, 00:10:25.130 "r_mbytes_per_sec": 0, 00:10:25.130 "w_mbytes_per_sec": 0 00:10:25.130 }, 00:10:25.130 "claimed": true, 00:10:25.130 "claim_type": "exclusive_write", 00:10:25.130 "zoned": false, 00:10:25.130 "supported_io_types": { 00:10:25.130 "read": true, 00:10:25.130 "write": true, 00:10:25.130 "unmap": true, 00:10:25.130 "flush": true, 00:10:25.130 "reset": true, 00:10:25.130 "nvme_admin": false, 00:10:25.130 "nvme_io": false, 00:10:25.130 "nvme_io_md": false, 00:10:25.130 "write_zeroes": true, 00:10:25.130 "zcopy": true, 00:10:25.130 "get_zone_info": false, 00:10:25.130 "zone_management": false, 00:10:25.130 "zone_append": false, 00:10:25.130 "compare": false, 00:10:25.130 "compare_and_write": false, 00:10:25.130 "abort": true, 00:10:25.130 "seek_hole": false, 00:10:25.130 "seek_data": false, 00:10:25.130 "copy": true, 00:10:25.130 "nvme_iov_md": false 00:10:25.130 }, 00:10:25.130 "memory_domains": [ 00:10:25.130 { 00:10:25.130 "dma_device_id": "system", 00:10:25.130 "dma_device_type": 1 00:10:25.130 }, 00:10:25.130 { 00:10:25.130 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:25.130 "dma_device_type": 2 00:10:25.130 } 00:10:25.130 ], 00:10:25.130 "driver_specific": {} 00:10:25.130 } 00:10:25.130 ] 00:10:25.130 15:16:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.130 15:16:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:25.130 15:16:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:25.130 15:16:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:25.130 15:16:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:10:25.130 15:16:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:25.130 15:16:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:25.130 15:16:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:25.130 15:16:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:25.130 15:16:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:25.130 15:16:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:25.130 15:16:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:25.130 15:16:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:25.130 15:16:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:25.130 15:16:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:25.130 15:16:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:25.130 15:16:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.130 15:16:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:25.130 15:16:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.130 15:16:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:25.130 "name": "Existed_Raid", 00:10:25.130 "uuid": "01327759-8873-4538-8fea-36c3305ff439", 00:10:25.130 "strip_size_kb": 0, 00:10:25.130 "state": "online", 00:10:25.130 "raid_level": "raid1", 00:10:25.130 "superblock": true, 00:10:25.130 "num_base_bdevs": 3, 00:10:25.130 "num_base_bdevs_discovered": 3, 00:10:25.130 "num_base_bdevs_operational": 3, 00:10:25.130 "base_bdevs_list": [ 00:10:25.130 { 00:10:25.130 "name": "BaseBdev1", 00:10:25.130 "uuid": "aee8f599-3136-49ab-b1fb-1d770643e893", 00:10:25.130 "is_configured": true, 00:10:25.130 "data_offset": 2048, 00:10:25.130 "data_size": 63488 00:10:25.130 }, 00:10:25.130 { 00:10:25.130 "name": "BaseBdev2", 00:10:25.130 "uuid": "168466db-c5dc-448f-a7ad-142fe4b607ca", 00:10:25.130 "is_configured": true, 00:10:25.130 "data_offset": 2048, 00:10:25.130 "data_size": 63488 00:10:25.130 }, 00:10:25.130 { 00:10:25.130 "name": "BaseBdev3", 00:10:25.130 "uuid": "005bfa81-fe96-4307-b30c-bfa6d8ac685c", 00:10:25.130 "is_configured": true, 00:10:25.130 "data_offset": 2048, 00:10:25.130 "data_size": 63488 00:10:25.130 } 00:10:25.130 ] 00:10:25.130 }' 00:10:25.130 15:16:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:25.394 15:16:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:25.653 15:16:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:10:25.653 15:16:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:25.653 15:16:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:25.653 15:16:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:25.653 15:16:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:10:25.653 15:16:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:25.653 15:16:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:25.653 15:16:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:25.653 15:16:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.653 15:16:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:25.653 [2024-11-19 15:16:15.893244] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:25.653 15:16:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.653 15:16:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:25.653 "name": "Existed_Raid", 00:10:25.653 "aliases": [ 00:10:25.653 "01327759-8873-4538-8fea-36c3305ff439" 00:10:25.653 ], 00:10:25.653 "product_name": "Raid Volume", 00:10:25.653 "block_size": 512, 00:10:25.653 "num_blocks": 63488, 00:10:25.653 "uuid": "01327759-8873-4538-8fea-36c3305ff439", 00:10:25.653 "assigned_rate_limits": { 00:10:25.653 "rw_ios_per_sec": 0, 00:10:25.653 "rw_mbytes_per_sec": 0, 00:10:25.653 "r_mbytes_per_sec": 0, 00:10:25.654 "w_mbytes_per_sec": 0 00:10:25.654 }, 00:10:25.654 "claimed": false, 00:10:25.654 "zoned": false, 00:10:25.654 "supported_io_types": { 00:10:25.654 "read": true, 00:10:25.654 "write": true, 00:10:25.654 "unmap": false, 00:10:25.654 "flush": false, 00:10:25.654 "reset": true, 00:10:25.654 "nvme_admin": false, 00:10:25.654 "nvme_io": false, 00:10:25.654 "nvme_io_md": false, 00:10:25.654 "write_zeroes": true, 00:10:25.654 "zcopy": false, 00:10:25.654 "get_zone_info": false, 00:10:25.654 "zone_management": false, 00:10:25.654 "zone_append": false, 00:10:25.654 "compare": false, 00:10:25.654 "compare_and_write": false, 00:10:25.654 "abort": false, 00:10:25.654 "seek_hole": false, 00:10:25.654 "seek_data": false, 00:10:25.654 "copy": false, 00:10:25.654 "nvme_iov_md": false 00:10:25.654 }, 00:10:25.654 "memory_domains": [ 00:10:25.654 { 00:10:25.654 "dma_device_id": "system", 00:10:25.654 "dma_device_type": 1 00:10:25.654 }, 00:10:25.654 { 00:10:25.654 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:25.654 "dma_device_type": 2 00:10:25.654 }, 00:10:25.654 { 00:10:25.654 "dma_device_id": "system", 00:10:25.654 "dma_device_type": 1 00:10:25.654 }, 00:10:25.654 { 00:10:25.654 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:25.654 "dma_device_type": 2 00:10:25.654 }, 00:10:25.654 { 00:10:25.654 "dma_device_id": "system", 00:10:25.654 "dma_device_type": 1 00:10:25.654 }, 00:10:25.654 { 00:10:25.654 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:25.654 "dma_device_type": 2 00:10:25.654 } 00:10:25.654 ], 00:10:25.654 "driver_specific": { 00:10:25.654 "raid": { 00:10:25.654 "uuid": "01327759-8873-4538-8fea-36c3305ff439", 00:10:25.654 "strip_size_kb": 0, 00:10:25.654 "state": "online", 00:10:25.654 "raid_level": "raid1", 00:10:25.654 "superblock": true, 00:10:25.654 "num_base_bdevs": 3, 00:10:25.654 "num_base_bdevs_discovered": 3, 00:10:25.654 "num_base_bdevs_operational": 3, 00:10:25.654 "base_bdevs_list": [ 00:10:25.654 { 00:10:25.654 "name": "BaseBdev1", 00:10:25.654 "uuid": "aee8f599-3136-49ab-b1fb-1d770643e893", 00:10:25.654 "is_configured": true, 00:10:25.654 "data_offset": 2048, 00:10:25.654 "data_size": 63488 00:10:25.654 }, 00:10:25.654 { 00:10:25.654 "name": "BaseBdev2", 00:10:25.654 "uuid": "168466db-c5dc-448f-a7ad-142fe4b607ca", 00:10:25.654 "is_configured": true, 00:10:25.654 "data_offset": 2048, 00:10:25.654 "data_size": 63488 00:10:25.654 }, 00:10:25.654 { 00:10:25.654 "name": "BaseBdev3", 00:10:25.654 "uuid": "005bfa81-fe96-4307-b30c-bfa6d8ac685c", 00:10:25.654 "is_configured": true, 00:10:25.654 "data_offset": 2048, 00:10:25.654 "data_size": 63488 00:10:25.654 } 00:10:25.654 ] 00:10:25.654 } 00:10:25.654 } 00:10:25.654 }' 00:10:25.654 15:16:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:25.654 15:16:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:10:25.654 BaseBdev2 00:10:25.654 BaseBdev3' 00:10:25.654 15:16:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:25.915 [2024-11-19 15:16:16.180371] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:25.915 "name": "Existed_Raid", 00:10:25.915 "uuid": "01327759-8873-4538-8fea-36c3305ff439", 00:10:25.915 "strip_size_kb": 0, 00:10:25.915 "state": "online", 00:10:25.915 "raid_level": "raid1", 00:10:25.915 "superblock": true, 00:10:25.915 "num_base_bdevs": 3, 00:10:25.915 "num_base_bdevs_discovered": 2, 00:10:25.915 "num_base_bdevs_operational": 2, 00:10:25.915 "base_bdevs_list": [ 00:10:25.915 { 00:10:25.915 "name": null, 00:10:25.915 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:25.915 "is_configured": false, 00:10:25.915 "data_offset": 0, 00:10:25.915 "data_size": 63488 00:10:25.915 }, 00:10:25.915 { 00:10:25.915 "name": "BaseBdev2", 00:10:25.915 "uuid": "168466db-c5dc-448f-a7ad-142fe4b607ca", 00:10:25.915 "is_configured": true, 00:10:25.915 "data_offset": 2048, 00:10:25.915 "data_size": 63488 00:10:25.915 }, 00:10:25.915 { 00:10:25.915 "name": "BaseBdev3", 00:10:25.915 "uuid": "005bfa81-fe96-4307-b30c-bfa6d8ac685c", 00:10:25.915 "is_configured": true, 00:10:25.915 "data_offset": 2048, 00:10:25.915 "data_size": 63488 00:10:25.915 } 00:10:25.915 ] 00:10:25.915 }' 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:25.915 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:26.485 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:10:26.485 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:26.485 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:26.485 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:26.485 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.485 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:26.485 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.485 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:26.485 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:26.485 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:10:26.485 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.485 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:26.485 [2024-11-19 15:16:16.680295] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:26.485 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.485 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:26.485 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:26.486 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:26.486 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.486 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:26.486 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:26.486 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.486 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:26.486 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:26.486 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:10:26.486 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.486 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:26.486 [2024-11-19 15:16:16.761011] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:26.486 [2024-11-19 15:16:16.761155] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:26.486 [2024-11-19 15:16:16.782263] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:26.486 [2024-11-19 15:16:16.782392] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:26.486 [2024-11-19 15:16:16.782450] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:10:26.486 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.486 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:26.486 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:26.486 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:26.486 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:10:26.486 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.486 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:26.486 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.746 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:10:26.746 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:26.747 BaseBdev2 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:26.747 [ 00:10:26.747 { 00:10:26.747 "name": "BaseBdev2", 00:10:26.747 "aliases": [ 00:10:26.747 "7cca44f7-4e2c-47d4-b016-9e1f2ae63ef8" 00:10:26.747 ], 00:10:26.747 "product_name": "Malloc disk", 00:10:26.747 "block_size": 512, 00:10:26.747 "num_blocks": 65536, 00:10:26.747 "uuid": "7cca44f7-4e2c-47d4-b016-9e1f2ae63ef8", 00:10:26.747 "assigned_rate_limits": { 00:10:26.747 "rw_ios_per_sec": 0, 00:10:26.747 "rw_mbytes_per_sec": 0, 00:10:26.747 "r_mbytes_per_sec": 0, 00:10:26.747 "w_mbytes_per_sec": 0 00:10:26.747 }, 00:10:26.747 "claimed": false, 00:10:26.747 "zoned": false, 00:10:26.747 "supported_io_types": { 00:10:26.747 "read": true, 00:10:26.747 "write": true, 00:10:26.747 "unmap": true, 00:10:26.747 "flush": true, 00:10:26.747 "reset": true, 00:10:26.747 "nvme_admin": false, 00:10:26.747 "nvme_io": false, 00:10:26.747 "nvme_io_md": false, 00:10:26.747 "write_zeroes": true, 00:10:26.747 "zcopy": true, 00:10:26.747 "get_zone_info": false, 00:10:26.747 "zone_management": false, 00:10:26.747 "zone_append": false, 00:10:26.747 "compare": false, 00:10:26.747 "compare_and_write": false, 00:10:26.747 "abort": true, 00:10:26.747 "seek_hole": false, 00:10:26.747 "seek_data": false, 00:10:26.747 "copy": true, 00:10:26.747 "nvme_iov_md": false 00:10:26.747 }, 00:10:26.747 "memory_domains": [ 00:10:26.747 { 00:10:26.747 "dma_device_id": "system", 00:10:26.747 "dma_device_type": 1 00:10:26.747 }, 00:10:26.747 { 00:10:26.747 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:26.747 "dma_device_type": 2 00:10:26.747 } 00:10:26.747 ], 00:10:26.747 "driver_specific": {} 00:10:26.747 } 00:10:26.747 ] 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:26.747 BaseBdev3 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:26.747 [ 00:10:26.747 { 00:10:26.747 "name": "BaseBdev3", 00:10:26.747 "aliases": [ 00:10:26.747 "ed339711-45a3-46a5-a8bf-63d2483d2fcd" 00:10:26.747 ], 00:10:26.747 "product_name": "Malloc disk", 00:10:26.747 "block_size": 512, 00:10:26.747 "num_blocks": 65536, 00:10:26.747 "uuid": "ed339711-45a3-46a5-a8bf-63d2483d2fcd", 00:10:26.747 "assigned_rate_limits": { 00:10:26.747 "rw_ios_per_sec": 0, 00:10:26.747 "rw_mbytes_per_sec": 0, 00:10:26.747 "r_mbytes_per_sec": 0, 00:10:26.747 "w_mbytes_per_sec": 0 00:10:26.747 }, 00:10:26.747 "claimed": false, 00:10:26.747 "zoned": false, 00:10:26.747 "supported_io_types": { 00:10:26.747 "read": true, 00:10:26.747 "write": true, 00:10:26.747 "unmap": true, 00:10:26.747 "flush": true, 00:10:26.747 "reset": true, 00:10:26.747 "nvme_admin": false, 00:10:26.747 "nvme_io": false, 00:10:26.747 "nvme_io_md": false, 00:10:26.747 "write_zeroes": true, 00:10:26.747 "zcopy": true, 00:10:26.747 "get_zone_info": false, 00:10:26.747 "zone_management": false, 00:10:26.747 "zone_append": false, 00:10:26.747 "compare": false, 00:10:26.747 "compare_and_write": false, 00:10:26.747 "abort": true, 00:10:26.747 "seek_hole": false, 00:10:26.747 "seek_data": false, 00:10:26.747 "copy": true, 00:10:26.747 "nvme_iov_md": false 00:10:26.747 }, 00:10:26.747 "memory_domains": [ 00:10:26.747 { 00:10:26.747 "dma_device_id": "system", 00:10:26.747 "dma_device_type": 1 00:10:26.747 }, 00:10:26.747 { 00:10:26.747 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:26.747 "dma_device_type": 2 00:10:26.747 } 00:10:26.747 ], 00:10:26.747 "driver_specific": {} 00:10:26.747 } 00:10:26.747 ] 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:26.747 [2024-11-19 15:16:16.957110] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:26.747 [2024-11-19 15:16:16.957245] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:26.747 [2024-11-19 15:16:16.957287] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:26.747 [2024-11-19 15:16:16.959584] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:26.747 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.748 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:26.748 15:16:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.748 15:16:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:26.748 "name": "Existed_Raid", 00:10:26.748 "uuid": "40c717fe-a639-4f46-94c4-341994c70c68", 00:10:26.748 "strip_size_kb": 0, 00:10:26.748 "state": "configuring", 00:10:26.748 "raid_level": "raid1", 00:10:26.748 "superblock": true, 00:10:26.748 "num_base_bdevs": 3, 00:10:26.748 "num_base_bdevs_discovered": 2, 00:10:26.748 "num_base_bdevs_operational": 3, 00:10:26.748 "base_bdevs_list": [ 00:10:26.748 { 00:10:26.748 "name": "BaseBdev1", 00:10:26.748 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:26.748 "is_configured": false, 00:10:26.748 "data_offset": 0, 00:10:26.748 "data_size": 0 00:10:26.748 }, 00:10:26.748 { 00:10:26.748 "name": "BaseBdev2", 00:10:26.748 "uuid": "7cca44f7-4e2c-47d4-b016-9e1f2ae63ef8", 00:10:26.748 "is_configured": true, 00:10:26.748 "data_offset": 2048, 00:10:26.748 "data_size": 63488 00:10:26.748 }, 00:10:26.748 { 00:10:26.748 "name": "BaseBdev3", 00:10:26.748 "uuid": "ed339711-45a3-46a5-a8bf-63d2483d2fcd", 00:10:26.748 "is_configured": true, 00:10:26.748 "data_offset": 2048, 00:10:26.748 "data_size": 63488 00:10:26.748 } 00:10:26.748 ] 00:10:26.748 }' 00:10:26.748 15:16:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:26.748 15:16:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:27.317 15:16:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:10:27.317 15:16:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.317 15:16:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:27.317 [2024-11-19 15:16:17.388426] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:27.317 15:16:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.317 15:16:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:10:27.317 15:16:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:27.317 15:16:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:27.317 15:16:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:27.317 15:16:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:27.317 15:16:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:27.317 15:16:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:27.317 15:16:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:27.317 15:16:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:27.317 15:16:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:27.318 15:16:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:27.318 15:16:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:27.318 15:16:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.318 15:16:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:27.318 15:16:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.318 15:16:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:27.318 "name": "Existed_Raid", 00:10:27.318 "uuid": "40c717fe-a639-4f46-94c4-341994c70c68", 00:10:27.318 "strip_size_kb": 0, 00:10:27.318 "state": "configuring", 00:10:27.318 "raid_level": "raid1", 00:10:27.318 "superblock": true, 00:10:27.318 "num_base_bdevs": 3, 00:10:27.318 "num_base_bdevs_discovered": 1, 00:10:27.318 "num_base_bdevs_operational": 3, 00:10:27.318 "base_bdevs_list": [ 00:10:27.318 { 00:10:27.318 "name": "BaseBdev1", 00:10:27.318 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:27.318 "is_configured": false, 00:10:27.318 "data_offset": 0, 00:10:27.318 "data_size": 0 00:10:27.318 }, 00:10:27.318 { 00:10:27.318 "name": null, 00:10:27.318 "uuid": "7cca44f7-4e2c-47d4-b016-9e1f2ae63ef8", 00:10:27.318 "is_configured": false, 00:10:27.318 "data_offset": 0, 00:10:27.318 "data_size": 63488 00:10:27.318 }, 00:10:27.318 { 00:10:27.318 "name": "BaseBdev3", 00:10:27.318 "uuid": "ed339711-45a3-46a5-a8bf-63d2483d2fcd", 00:10:27.318 "is_configured": true, 00:10:27.318 "data_offset": 2048, 00:10:27.318 "data_size": 63488 00:10:27.318 } 00:10:27.318 ] 00:10:27.318 }' 00:10:27.318 15:16:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:27.318 15:16:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:27.578 15:16:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:27.578 15:16:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.578 15:16:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:27.578 15:16:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:27.578 15:16:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.578 15:16:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:10:27.578 15:16:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:27.578 15:16:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.578 15:16:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:27.578 [2024-11-19 15:16:17.876379] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:27.578 BaseBdev1 00:10:27.578 15:16:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.578 15:16:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:10:27.578 15:16:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:10:27.578 15:16:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:27.578 15:16:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:27.578 15:16:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:27.578 15:16:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:27.578 15:16:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:27.578 15:16:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.578 15:16:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:27.578 15:16:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.578 15:16:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:27.578 15:16:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.578 15:16:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:27.578 [ 00:10:27.578 { 00:10:27.578 "name": "BaseBdev1", 00:10:27.578 "aliases": [ 00:10:27.578 "0251798a-793f-4ee4-93a4-d2a5ed21be64" 00:10:27.578 ], 00:10:27.578 "product_name": "Malloc disk", 00:10:27.578 "block_size": 512, 00:10:27.578 "num_blocks": 65536, 00:10:27.578 "uuid": "0251798a-793f-4ee4-93a4-d2a5ed21be64", 00:10:27.578 "assigned_rate_limits": { 00:10:27.578 "rw_ios_per_sec": 0, 00:10:27.578 "rw_mbytes_per_sec": 0, 00:10:27.578 "r_mbytes_per_sec": 0, 00:10:27.578 "w_mbytes_per_sec": 0 00:10:27.578 }, 00:10:27.578 "claimed": true, 00:10:27.578 "claim_type": "exclusive_write", 00:10:27.578 "zoned": false, 00:10:27.578 "supported_io_types": { 00:10:27.578 "read": true, 00:10:27.578 "write": true, 00:10:27.578 "unmap": true, 00:10:27.578 "flush": true, 00:10:27.578 "reset": true, 00:10:27.578 "nvme_admin": false, 00:10:27.578 "nvme_io": false, 00:10:27.578 "nvme_io_md": false, 00:10:27.578 "write_zeroes": true, 00:10:27.578 "zcopy": true, 00:10:27.578 "get_zone_info": false, 00:10:27.578 "zone_management": false, 00:10:27.578 "zone_append": false, 00:10:27.578 "compare": false, 00:10:27.578 "compare_and_write": false, 00:10:27.578 "abort": true, 00:10:27.578 "seek_hole": false, 00:10:27.578 "seek_data": false, 00:10:27.578 "copy": true, 00:10:27.578 "nvme_iov_md": false 00:10:27.578 }, 00:10:27.578 "memory_domains": [ 00:10:27.578 { 00:10:27.578 "dma_device_id": "system", 00:10:27.578 "dma_device_type": 1 00:10:27.578 }, 00:10:27.578 { 00:10:27.578 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:27.578 "dma_device_type": 2 00:10:27.578 } 00:10:27.578 ], 00:10:27.578 "driver_specific": {} 00:10:27.578 } 00:10:27.578 ] 00:10:27.578 15:16:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.578 15:16:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:27.578 15:16:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:10:27.578 15:16:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:27.839 15:16:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:27.839 15:16:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:27.839 15:16:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:27.839 15:16:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:27.839 15:16:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:27.839 15:16:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:27.839 15:16:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:27.839 15:16:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:27.839 15:16:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:27.839 15:16:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:27.839 15:16:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:27.839 15:16:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:27.839 15:16:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:27.839 15:16:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:27.839 "name": "Existed_Raid", 00:10:27.839 "uuid": "40c717fe-a639-4f46-94c4-341994c70c68", 00:10:27.839 "strip_size_kb": 0, 00:10:27.839 "state": "configuring", 00:10:27.839 "raid_level": "raid1", 00:10:27.839 "superblock": true, 00:10:27.839 "num_base_bdevs": 3, 00:10:27.839 "num_base_bdevs_discovered": 2, 00:10:27.839 "num_base_bdevs_operational": 3, 00:10:27.839 "base_bdevs_list": [ 00:10:27.839 { 00:10:27.839 "name": "BaseBdev1", 00:10:27.839 "uuid": "0251798a-793f-4ee4-93a4-d2a5ed21be64", 00:10:27.839 "is_configured": true, 00:10:27.839 "data_offset": 2048, 00:10:27.839 "data_size": 63488 00:10:27.839 }, 00:10:27.839 { 00:10:27.839 "name": null, 00:10:27.839 "uuid": "7cca44f7-4e2c-47d4-b016-9e1f2ae63ef8", 00:10:27.839 "is_configured": false, 00:10:27.839 "data_offset": 0, 00:10:27.839 "data_size": 63488 00:10:27.839 }, 00:10:27.839 { 00:10:27.839 "name": "BaseBdev3", 00:10:27.839 "uuid": "ed339711-45a3-46a5-a8bf-63d2483d2fcd", 00:10:27.839 "is_configured": true, 00:10:27.839 "data_offset": 2048, 00:10:27.839 "data_size": 63488 00:10:27.839 } 00:10:27.839 ] 00:10:27.839 }' 00:10:27.839 15:16:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:27.839 15:16:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:28.099 15:16:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:28.099 15:16:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:28.099 15:16:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.099 15:16:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:28.099 15:16:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.099 15:16:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:10:28.099 15:16:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:10:28.099 15:16:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.099 15:16:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:28.099 [2024-11-19 15:16:18.407703] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:28.099 15:16:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.099 15:16:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:10:28.099 15:16:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:28.099 15:16:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:28.099 15:16:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:28.099 15:16:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:28.099 15:16:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:28.099 15:16:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:28.099 15:16:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:28.099 15:16:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:28.099 15:16:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:28.099 15:16:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:28.099 15:16:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.099 15:16:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:28.099 15:16:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:28.099 15:16:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.358 15:16:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:28.358 "name": "Existed_Raid", 00:10:28.358 "uuid": "40c717fe-a639-4f46-94c4-341994c70c68", 00:10:28.358 "strip_size_kb": 0, 00:10:28.358 "state": "configuring", 00:10:28.359 "raid_level": "raid1", 00:10:28.359 "superblock": true, 00:10:28.359 "num_base_bdevs": 3, 00:10:28.359 "num_base_bdevs_discovered": 1, 00:10:28.359 "num_base_bdevs_operational": 3, 00:10:28.359 "base_bdevs_list": [ 00:10:28.359 { 00:10:28.359 "name": "BaseBdev1", 00:10:28.359 "uuid": "0251798a-793f-4ee4-93a4-d2a5ed21be64", 00:10:28.359 "is_configured": true, 00:10:28.359 "data_offset": 2048, 00:10:28.359 "data_size": 63488 00:10:28.359 }, 00:10:28.359 { 00:10:28.359 "name": null, 00:10:28.359 "uuid": "7cca44f7-4e2c-47d4-b016-9e1f2ae63ef8", 00:10:28.359 "is_configured": false, 00:10:28.359 "data_offset": 0, 00:10:28.359 "data_size": 63488 00:10:28.359 }, 00:10:28.359 { 00:10:28.359 "name": null, 00:10:28.359 "uuid": "ed339711-45a3-46a5-a8bf-63d2483d2fcd", 00:10:28.359 "is_configured": false, 00:10:28.359 "data_offset": 0, 00:10:28.359 "data_size": 63488 00:10:28.359 } 00:10:28.359 ] 00:10:28.359 }' 00:10:28.359 15:16:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:28.359 15:16:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:28.619 15:16:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:28.619 15:16:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:28.619 15:16:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.619 15:16:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:28.619 15:16:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.619 15:16:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:10:28.619 15:16:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:10:28.619 15:16:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.619 15:16:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:28.619 [2024-11-19 15:16:18.854921] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:28.619 15:16:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.619 15:16:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:10:28.619 15:16:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:28.619 15:16:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:28.619 15:16:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:28.619 15:16:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:28.619 15:16:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:28.619 15:16:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:28.619 15:16:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:28.619 15:16:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:28.619 15:16:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:28.619 15:16:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:28.619 15:16:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:28.619 15:16:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.619 15:16:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:28.619 15:16:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.619 15:16:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:28.619 "name": "Existed_Raid", 00:10:28.619 "uuid": "40c717fe-a639-4f46-94c4-341994c70c68", 00:10:28.619 "strip_size_kb": 0, 00:10:28.619 "state": "configuring", 00:10:28.619 "raid_level": "raid1", 00:10:28.619 "superblock": true, 00:10:28.619 "num_base_bdevs": 3, 00:10:28.619 "num_base_bdevs_discovered": 2, 00:10:28.619 "num_base_bdevs_operational": 3, 00:10:28.619 "base_bdevs_list": [ 00:10:28.619 { 00:10:28.619 "name": "BaseBdev1", 00:10:28.619 "uuid": "0251798a-793f-4ee4-93a4-d2a5ed21be64", 00:10:28.619 "is_configured": true, 00:10:28.619 "data_offset": 2048, 00:10:28.619 "data_size": 63488 00:10:28.619 }, 00:10:28.619 { 00:10:28.619 "name": null, 00:10:28.619 "uuid": "7cca44f7-4e2c-47d4-b016-9e1f2ae63ef8", 00:10:28.619 "is_configured": false, 00:10:28.619 "data_offset": 0, 00:10:28.619 "data_size": 63488 00:10:28.619 }, 00:10:28.619 { 00:10:28.619 "name": "BaseBdev3", 00:10:28.619 "uuid": "ed339711-45a3-46a5-a8bf-63d2483d2fcd", 00:10:28.619 "is_configured": true, 00:10:28.619 "data_offset": 2048, 00:10:28.619 "data_size": 63488 00:10:28.619 } 00:10:28.619 ] 00:10:28.619 }' 00:10:28.619 15:16:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:28.619 15:16:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:29.188 15:16:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:29.188 15:16:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:29.188 15:16:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:29.188 15:16:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:29.188 15:16:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:29.188 15:16:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:10:29.188 15:16:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:29.188 15:16:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:29.188 15:16:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:29.188 [2024-11-19 15:16:19.290215] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:29.188 15:16:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:29.188 15:16:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:10:29.188 15:16:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:29.188 15:16:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:29.188 15:16:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:29.188 15:16:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:29.188 15:16:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:29.188 15:16:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:29.188 15:16:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:29.188 15:16:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:29.188 15:16:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:29.188 15:16:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:29.188 15:16:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:29.188 15:16:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:29.188 15:16:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:29.188 15:16:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:29.188 15:16:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:29.188 "name": "Existed_Raid", 00:10:29.188 "uuid": "40c717fe-a639-4f46-94c4-341994c70c68", 00:10:29.188 "strip_size_kb": 0, 00:10:29.188 "state": "configuring", 00:10:29.188 "raid_level": "raid1", 00:10:29.188 "superblock": true, 00:10:29.188 "num_base_bdevs": 3, 00:10:29.188 "num_base_bdevs_discovered": 1, 00:10:29.188 "num_base_bdevs_operational": 3, 00:10:29.188 "base_bdevs_list": [ 00:10:29.188 { 00:10:29.188 "name": null, 00:10:29.188 "uuid": "0251798a-793f-4ee4-93a4-d2a5ed21be64", 00:10:29.188 "is_configured": false, 00:10:29.188 "data_offset": 0, 00:10:29.188 "data_size": 63488 00:10:29.188 }, 00:10:29.188 { 00:10:29.188 "name": null, 00:10:29.188 "uuid": "7cca44f7-4e2c-47d4-b016-9e1f2ae63ef8", 00:10:29.188 "is_configured": false, 00:10:29.188 "data_offset": 0, 00:10:29.188 "data_size": 63488 00:10:29.188 }, 00:10:29.188 { 00:10:29.188 "name": "BaseBdev3", 00:10:29.188 "uuid": "ed339711-45a3-46a5-a8bf-63d2483d2fcd", 00:10:29.188 "is_configured": true, 00:10:29.188 "data_offset": 2048, 00:10:29.188 "data_size": 63488 00:10:29.188 } 00:10:29.188 ] 00:10:29.188 }' 00:10:29.188 15:16:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:29.188 15:16:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:29.448 15:16:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:29.448 15:16:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:29.448 15:16:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:29.448 15:16:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:29.448 15:16:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:29.448 15:16:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:10:29.448 15:16:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:10:29.448 15:16:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:29.448 15:16:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:29.448 [2024-11-19 15:16:19.761410] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:29.448 15:16:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:29.448 15:16:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:10:29.448 15:16:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:29.448 15:16:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:29.448 15:16:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:29.448 15:16:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:29.448 15:16:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:29.448 15:16:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:29.448 15:16:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:29.448 15:16:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:29.448 15:16:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:29.448 15:16:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:29.448 15:16:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:29.448 15:16:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:29.448 15:16:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:29.448 15:16:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:29.708 15:16:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:29.708 "name": "Existed_Raid", 00:10:29.708 "uuid": "40c717fe-a639-4f46-94c4-341994c70c68", 00:10:29.708 "strip_size_kb": 0, 00:10:29.708 "state": "configuring", 00:10:29.708 "raid_level": "raid1", 00:10:29.708 "superblock": true, 00:10:29.708 "num_base_bdevs": 3, 00:10:29.708 "num_base_bdevs_discovered": 2, 00:10:29.708 "num_base_bdevs_operational": 3, 00:10:29.708 "base_bdevs_list": [ 00:10:29.708 { 00:10:29.708 "name": null, 00:10:29.708 "uuid": "0251798a-793f-4ee4-93a4-d2a5ed21be64", 00:10:29.708 "is_configured": false, 00:10:29.708 "data_offset": 0, 00:10:29.708 "data_size": 63488 00:10:29.708 }, 00:10:29.708 { 00:10:29.708 "name": "BaseBdev2", 00:10:29.708 "uuid": "7cca44f7-4e2c-47d4-b016-9e1f2ae63ef8", 00:10:29.708 "is_configured": true, 00:10:29.708 "data_offset": 2048, 00:10:29.708 "data_size": 63488 00:10:29.708 }, 00:10:29.708 { 00:10:29.708 "name": "BaseBdev3", 00:10:29.708 "uuid": "ed339711-45a3-46a5-a8bf-63d2483d2fcd", 00:10:29.708 "is_configured": true, 00:10:29.708 "data_offset": 2048, 00:10:29.708 "data_size": 63488 00:10:29.708 } 00:10:29.708 ] 00:10:29.708 }' 00:10:29.708 15:16:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:29.708 15:16:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:29.967 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:29.967 15:16:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:29.967 15:16:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:29.967 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:29.967 15:16:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:29.967 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:10:29.967 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:10:29.967 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:29.967 15:16:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:29.967 15:16:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:29.967 15:16:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:29.968 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 0251798a-793f-4ee4-93a4-d2a5ed21be64 00:10:29.968 15:16:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:29.968 15:16:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:30.227 NewBaseBdev 00:10:30.227 [2024-11-19 15:16:20.317241] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:10:30.227 [2024-11-19 15:16:20.317439] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:10:30.227 [2024-11-19 15:16:20.317453] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:30.227 [2024-11-19 15:16:20.317737] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:10:30.227 [2024-11-19 15:16:20.317857] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:10:30.227 [2024-11-19 15:16:20.317871] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:10:30.227 [2024-11-19 15:16:20.318004] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:30.227 15:16:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:30.227 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:10:30.227 15:16:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:10:30.227 15:16:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:30.227 15:16:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:30.227 15:16:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:30.227 15:16:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:30.227 15:16:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:30.227 15:16:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:30.227 15:16:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:30.227 15:16:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:30.227 15:16:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:10:30.227 15:16:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:30.227 15:16:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:30.227 [ 00:10:30.227 { 00:10:30.227 "name": "NewBaseBdev", 00:10:30.227 "aliases": [ 00:10:30.227 "0251798a-793f-4ee4-93a4-d2a5ed21be64" 00:10:30.227 ], 00:10:30.227 "product_name": "Malloc disk", 00:10:30.227 "block_size": 512, 00:10:30.227 "num_blocks": 65536, 00:10:30.227 "uuid": "0251798a-793f-4ee4-93a4-d2a5ed21be64", 00:10:30.227 "assigned_rate_limits": { 00:10:30.227 "rw_ios_per_sec": 0, 00:10:30.227 "rw_mbytes_per_sec": 0, 00:10:30.227 "r_mbytes_per_sec": 0, 00:10:30.227 "w_mbytes_per_sec": 0 00:10:30.227 }, 00:10:30.227 "claimed": true, 00:10:30.227 "claim_type": "exclusive_write", 00:10:30.227 "zoned": false, 00:10:30.227 "supported_io_types": { 00:10:30.227 "read": true, 00:10:30.227 "write": true, 00:10:30.227 "unmap": true, 00:10:30.227 "flush": true, 00:10:30.227 "reset": true, 00:10:30.227 "nvme_admin": false, 00:10:30.227 "nvme_io": false, 00:10:30.227 "nvme_io_md": false, 00:10:30.227 "write_zeroes": true, 00:10:30.227 "zcopy": true, 00:10:30.227 "get_zone_info": false, 00:10:30.227 "zone_management": false, 00:10:30.227 "zone_append": false, 00:10:30.227 "compare": false, 00:10:30.227 "compare_and_write": false, 00:10:30.227 "abort": true, 00:10:30.227 "seek_hole": false, 00:10:30.227 "seek_data": false, 00:10:30.227 "copy": true, 00:10:30.227 "nvme_iov_md": false 00:10:30.227 }, 00:10:30.227 "memory_domains": [ 00:10:30.227 { 00:10:30.227 "dma_device_id": "system", 00:10:30.227 "dma_device_type": 1 00:10:30.227 }, 00:10:30.227 { 00:10:30.227 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:30.227 "dma_device_type": 2 00:10:30.227 } 00:10:30.227 ], 00:10:30.227 "driver_specific": {} 00:10:30.227 } 00:10:30.227 ] 00:10:30.227 15:16:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:30.227 15:16:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:30.227 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:10:30.227 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:30.227 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:30.227 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:30.227 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:30.227 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:30.227 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:30.227 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:30.227 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:30.227 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:30.227 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:30.227 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:30.227 15:16:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:30.227 15:16:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:30.227 15:16:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:30.228 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:30.228 "name": "Existed_Raid", 00:10:30.228 "uuid": "40c717fe-a639-4f46-94c4-341994c70c68", 00:10:30.228 "strip_size_kb": 0, 00:10:30.228 "state": "online", 00:10:30.228 "raid_level": "raid1", 00:10:30.228 "superblock": true, 00:10:30.228 "num_base_bdevs": 3, 00:10:30.228 "num_base_bdevs_discovered": 3, 00:10:30.228 "num_base_bdevs_operational": 3, 00:10:30.228 "base_bdevs_list": [ 00:10:30.228 { 00:10:30.228 "name": "NewBaseBdev", 00:10:30.228 "uuid": "0251798a-793f-4ee4-93a4-d2a5ed21be64", 00:10:30.228 "is_configured": true, 00:10:30.228 "data_offset": 2048, 00:10:30.228 "data_size": 63488 00:10:30.228 }, 00:10:30.228 { 00:10:30.228 "name": "BaseBdev2", 00:10:30.228 "uuid": "7cca44f7-4e2c-47d4-b016-9e1f2ae63ef8", 00:10:30.228 "is_configured": true, 00:10:30.228 "data_offset": 2048, 00:10:30.228 "data_size": 63488 00:10:30.228 }, 00:10:30.228 { 00:10:30.228 "name": "BaseBdev3", 00:10:30.228 "uuid": "ed339711-45a3-46a5-a8bf-63d2483d2fcd", 00:10:30.228 "is_configured": true, 00:10:30.228 "data_offset": 2048, 00:10:30.228 "data_size": 63488 00:10:30.228 } 00:10:30.228 ] 00:10:30.228 }' 00:10:30.228 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:30.228 15:16:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:30.488 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:10:30.488 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:30.488 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:30.488 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:30.488 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:10:30.488 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:30.488 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:30.488 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:30.488 15:16:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:30.488 15:16:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:30.488 [2024-11-19 15:16:20.796813] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:30.488 15:16:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:30.488 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:30.488 "name": "Existed_Raid", 00:10:30.488 "aliases": [ 00:10:30.488 "40c717fe-a639-4f46-94c4-341994c70c68" 00:10:30.488 ], 00:10:30.488 "product_name": "Raid Volume", 00:10:30.488 "block_size": 512, 00:10:30.488 "num_blocks": 63488, 00:10:30.488 "uuid": "40c717fe-a639-4f46-94c4-341994c70c68", 00:10:30.488 "assigned_rate_limits": { 00:10:30.488 "rw_ios_per_sec": 0, 00:10:30.488 "rw_mbytes_per_sec": 0, 00:10:30.488 "r_mbytes_per_sec": 0, 00:10:30.488 "w_mbytes_per_sec": 0 00:10:30.488 }, 00:10:30.488 "claimed": false, 00:10:30.488 "zoned": false, 00:10:30.488 "supported_io_types": { 00:10:30.488 "read": true, 00:10:30.488 "write": true, 00:10:30.488 "unmap": false, 00:10:30.488 "flush": false, 00:10:30.488 "reset": true, 00:10:30.488 "nvme_admin": false, 00:10:30.488 "nvme_io": false, 00:10:30.488 "nvme_io_md": false, 00:10:30.488 "write_zeroes": true, 00:10:30.488 "zcopy": false, 00:10:30.488 "get_zone_info": false, 00:10:30.488 "zone_management": false, 00:10:30.488 "zone_append": false, 00:10:30.488 "compare": false, 00:10:30.488 "compare_and_write": false, 00:10:30.488 "abort": false, 00:10:30.488 "seek_hole": false, 00:10:30.488 "seek_data": false, 00:10:30.488 "copy": false, 00:10:30.488 "nvme_iov_md": false 00:10:30.488 }, 00:10:30.488 "memory_domains": [ 00:10:30.488 { 00:10:30.488 "dma_device_id": "system", 00:10:30.488 "dma_device_type": 1 00:10:30.488 }, 00:10:30.488 { 00:10:30.488 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:30.488 "dma_device_type": 2 00:10:30.488 }, 00:10:30.488 { 00:10:30.488 "dma_device_id": "system", 00:10:30.488 "dma_device_type": 1 00:10:30.488 }, 00:10:30.488 { 00:10:30.488 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:30.488 "dma_device_type": 2 00:10:30.488 }, 00:10:30.488 { 00:10:30.488 "dma_device_id": "system", 00:10:30.488 "dma_device_type": 1 00:10:30.488 }, 00:10:30.488 { 00:10:30.488 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:30.488 "dma_device_type": 2 00:10:30.488 } 00:10:30.488 ], 00:10:30.488 "driver_specific": { 00:10:30.488 "raid": { 00:10:30.488 "uuid": "40c717fe-a639-4f46-94c4-341994c70c68", 00:10:30.488 "strip_size_kb": 0, 00:10:30.488 "state": "online", 00:10:30.488 "raid_level": "raid1", 00:10:30.488 "superblock": true, 00:10:30.488 "num_base_bdevs": 3, 00:10:30.488 "num_base_bdevs_discovered": 3, 00:10:30.488 "num_base_bdevs_operational": 3, 00:10:30.488 "base_bdevs_list": [ 00:10:30.488 { 00:10:30.488 "name": "NewBaseBdev", 00:10:30.488 "uuid": "0251798a-793f-4ee4-93a4-d2a5ed21be64", 00:10:30.488 "is_configured": true, 00:10:30.488 "data_offset": 2048, 00:10:30.488 "data_size": 63488 00:10:30.488 }, 00:10:30.488 { 00:10:30.488 "name": "BaseBdev2", 00:10:30.488 "uuid": "7cca44f7-4e2c-47d4-b016-9e1f2ae63ef8", 00:10:30.488 "is_configured": true, 00:10:30.488 "data_offset": 2048, 00:10:30.488 "data_size": 63488 00:10:30.488 }, 00:10:30.488 { 00:10:30.488 "name": "BaseBdev3", 00:10:30.488 "uuid": "ed339711-45a3-46a5-a8bf-63d2483d2fcd", 00:10:30.488 "is_configured": true, 00:10:30.488 "data_offset": 2048, 00:10:30.488 "data_size": 63488 00:10:30.488 } 00:10:30.488 ] 00:10:30.488 } 00:10:30.488 } 00:10:30.488 }' 00:10:30.488 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:30.749 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:10:30.749 BaseBdev2 00:10:30.749 BaseBdev3' 00:10:30.749 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:30.749 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:30.749 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:30.749 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:10:30.749 15:16:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:30.749 15:16:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:30.749 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:30.749 15:16:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:30.749 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:30.749 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:30.749 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:30.749 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:30.749 15:16:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:30.749 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:30.749 15:16:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:30.749 15:16:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:30.749 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:30.749 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:30.749 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:30.749 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:30.749 15:16:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:30.749 15:16:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:30.749 15:16:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:30.749 15:16:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:30.749 15:16:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:30.749 15:16:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:30.749 15:16:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:30.749 15:16:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:30.749 15:16:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:30.749 [2024-11-19 15:16:21.044043] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:30.749 [2024-11-19 15:16:21.044086] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:30.749 [2024-11-19 15:16:21.044166] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:30.749 [2024-11-19 15:16:21.044441] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:30.749 [2024-11-19 15:16:21.044452] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:10:30.749 15:16:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:30.749 15:16:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 79053 00:10:30.749 15:16:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 79053 ']' 00:10:30.749 15:16:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 79053 00:10:30.749 15:16:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:10:30.749 15:16:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:30.749 15:16:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 79053 00:10:30.749 15:16:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:30.749 15:16:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:30.749 15:16:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 79053' 00:10:30.749 killing process with pid 79053 00:10:30.749 15:16:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 79053 00:10:30.749 [2024-11-19 15:16:21.084207] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:30.749 15:16:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 79053 00:10:31.009 [2024-11-19 15:16:21.143153] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:31.269 15:16:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:10:31.269 00:10:31.269 real 0m8.982s 00:10:31.269 user 0m15.099s 00:10:31.269 sys 0m1.867s 00:10:31.269 15:16:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:31.269 15:16:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:31.269 ************************************ 00:10:31.269 END TEST raid_state_function_test_sb 00:10:31.269 ************************************ 00:10:31.269 15:16:21 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid1 3 00:10:31.269 15:16:21 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:10:31.269 15:16:21 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:31.269 15:16:21 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:31.269 ************************************ 00:10:31.269 START TEST raid_superblock_test 00:10:31.269 ************************************ 00:10:31.269 15:16:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid1 3 00:10:31.269 15:16:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:10:31.269 15:16:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:10:31.269 15:16:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:10:31.269 15:16:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:10:31.269 15:16:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:10:31.269 15:16:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:10:31.269 15:16:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:10:31.269 15:16:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:10:31.269 15:16:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:10:31.269 15:16:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:10:31.269 15:16:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:10:31.269 15:16:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:10:31.269 15:16:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:10:31.269 15:16:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:10:31.269 15:16:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:10:31.269 15:16:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=79662 00:10:31.269 15:16:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:10:31.269 15:16:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 79662 00:10:31.269 15:16:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 79662 ']' 00:10:31.269 15:16:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:31.269 15:16:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:31.269 15:16:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:31.269 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:31.269 15:16:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:31.269 15:16:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:31.529 [2024-11-19 15:16:21.626233] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:10:31.529 [2024-11-19 15:16:21.626430] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid79662 ] 00:10:31.529 [2024-11-19 15:16:21.781041] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:31.529 [2024-11-19 15:16:21.822351] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:31.788 [2024-11-19 15:16:21.900303] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:31.788 [2024-11-19 15:16:21.900448] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:32.358 15:16:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:32.358 15:16:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:10:32.358 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:10:32.358 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:32.358 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:10:32.358 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:10:32.358 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:10:32.358 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:32.358 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:32.358 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:32.358 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:10:32.358 15:16:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:32.358 15:16:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.358 malloc1 00:10:32.358 15:16:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:32.358 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:32.358 15:16:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:32.358 15:16:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.358 [2024-11-19 15:16:22.522999] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:32.358 [2024-11-19 15:16:22.523171] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:32.358 [2024-11-19 15:16:22.523209] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:10:32.358 [2024-11-19 15:16:22.523256] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:32.358 [2024-11-19 15:16:22.525726] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:32.358 [2024-11-19 15:16:22.525823] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:32.358 pt1 00:10:32.358 15:16:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:32.358 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:32.358 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:32.358 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:10:32.358 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:10:32.358 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:10:32.358 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:32.358 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:32.358 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:32.358 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:10:32.358 15:16:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:32.358 15:16:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.358 malloc2 00:10:32.358 15:16:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:32.359 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:32.359 15:16:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:32.359 15:16:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.359 [2024-11-19 15:16:22.561462] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:32.359 [2024-11-19 15:16:22.561579] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:32.359 [2024-11-19 15:16:22.561612] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:10:32.359 [2024-11-19 15:16:22.561642] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:32.359 [2024-11-19 15:16:22.564032] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:32.359 [2024-11-19 15:16:22.564105] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:32.359 pt2 00:10:32.359 15:16:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:32.359 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:32.359 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:32.359 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:10:32.359 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:10:32.359 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:10:32.359 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:32.359 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:32.359 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:32.359 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:10:32.359 15:16:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:32.359 15:16:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.359 malloc3 00:10:32.359 15:16:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:32.359 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:32.359 15:16:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:32.359 15:16:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.359 [2024-11-19 15:16:22.599840] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:32.359 [2024-11-19 15:16:22.599963] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:32.359 [2024-11-19 15:16:22.600016] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:10:32.359 [2024-11-19 15:16:22.600052] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:32.359 [2024-11-19 15:16:22.602390] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:32.359 [2024-11-19 15:16:22.602462] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:32.359 pt3 00:10:32.359 15:16:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:32.359 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:32.359 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:32.359 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:10:32.359 15:16:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:32.359 15:16:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.359 [2024-11-19 15:16:22.611882] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:32.359 [2024-11-19 15:16:22.614038] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:32.359 [2024-11-19 15:16:22.614132] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:32.359 [2024-11-19 15:16:22.614316] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:10:32.359 [2024-11-19 15:16:22.614363] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:32.359 [2024-11-19 15:16:22.614643] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:10:32.359 [2024-11-19 15:16:22.614824] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:10:32.359 [2024-11-19 15:16:22.614877] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:10:32.359 [2024-11-19 15:16:22.615050] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:32.359 15:16:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:32.359 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:10:32.359 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:32.359 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:32.359 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:32.359 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:32.359 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:32.359 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:32.359 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:32.359 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:32.359 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:32.359 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:32.359 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:32.359 15:16:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:32.359 15:16:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.359 15:16:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:32.359 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:32.359 "name": "raid_bdev1", 00:10:32.359 "uuid": "6faddf11-01b1-41c9-add2-fa8f648df3e4", 00:10:32.359 "strip_size_kb": 0, 00:10:32.359 "state": "online", 00:10:32.359 "raid_level": "raid1", 00:10:32.359 "superblock": true, 00:10:32.359 "num_base_bdevs": 3, 00:10:32.359 "num_base_bdevs_discovered": 3, 00:10:32.359 "num_base_bdevs_operational": 3, 00:10:32.359 "base_bdevs_list": [ 00:10:32.359 { 00:10:32.359 "name": "pt1", 00:10:32.359 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:32.359 "is_configured": true, 00:10:32.359 "data_offset": 2048, 00:10:32.359 "data_size": 63488 00:10:32.359 }, 00:10:32.359 { 00:10:32.359 "name": "pt2", 00:10:32.359 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:32.359 "is_configured": true, 00:10:32.359 "data_offset": 2048, 00:10:32.359 "data_size": 63488 00:10:32.359 }, 00:10:32.359 { 00:10:32.359 "name": "pt3", 00:10:32.359 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:32.359 "is_configured": true, 00:10:32.359 "data_offset": 2048, 00:10:32.359 "data_size": 63488 00:10:32.359 } 00:10:32.359 ] 00:10:32.359 }' 00:10:32.359 15:16:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:32.359 15:16:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.928 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:10:32.929 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:10:32.929 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:32.929 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:32.929 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:32.929 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:32.929 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:32.929 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:32.929 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.929 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:32.929 [2024-11-19 15:16:23.087373] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:32.929 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:32.929 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:32.929 "name": "raid_bdev1", 00:10:32.929 "aliases": [ 00:10:32.929 "6faddf11-01b1-41c9-add2-fa8f648df3e4" 00:10:32.929 ], 00:10:32.929 "product_name": "Raid Volume", 00:10:32.929 "block_size": 512, 00:10:32.929 "num_blocks": 63488, 00:10:32.929 "uuid": "6faddf11-01b1-41c9-add2-fa8f648df3e4", 00:10:32.929 "assigned_rate_limits": { 00:10:32.929 "rw_ios_per_sec": 0, 00:10:32.929 "rw_mbytes_per_sec": 0, 00:10:32.929 "r_mbytes_per_sec": 0, 00:10:32.929 "w_mbytes_per_sec": 0 00:10:32.929 }, 00:10:32.929 "claimed": false, 00:10:32.929 "zoned": false, 00:10:32.929 "supported_io_types": { 00:10:32.929 "read": true, 00:10:32.929 "write": true, 00:10:32.929 "unmap": false, 00:10:32.929 "flush": false, 00:10:32.929 "reset": true, 00:10:32.929 "nvme_admin": false, 00:10:32.929 "nvme_io": false, 00:10:32.929 "nvme_io_md": false, 00:10:32.929 "write_zeroes": true, 00:10:32.929 "zcopy": false, 00:10:32.929 "get_zone_info": false, 00:10:32.929 "zone_management": false, 00:10:32.929 "zone_append": false, 00:10:32.929 "compare": false, 00:10:32.929 "compare_and_write": false, 00:10:32.929 "abort": false, 00:10:32.929 "seek_hole": false, 00:10:32.929 "seek_data": false, 00:10:32.929 "copy": false, 00:10:32.929 "nvme_iov_md": false 00:10:32.929 }, 00:10:32.929 "memory_domains": [ 00:10:32.929 { 00:10:32.929 "dma_device_id": "system", 00:10:32.929 "dma_device_type": 1 00:10:32.929 }, 00:10:32.929 { 00:10:32.929 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:32.929 "dma_device_type": 2 00:10:32.929 }, 00:10:32.929 { 00:10:32.929 "dma_device_id": "system", 00:10:32.929 "dma_device_type": 1 00:10:32.929 }, 00:10:32.929 { 00:10:32.929 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:32.929 "dma_device_type": 2 00:10:32.929 }, 00:10:32.929 { 00:10:32.929 "dma_device_id": "system", 00:10:32.929 "dma_device_type": 1 00:10:32.929 }, 00:10:32.929 { 00:10:32.929 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:32.929 "dma_device_type": 2 00:10:32.929 } 00:10:32.929 ], 00:10:32.929 "driver_specific": { 00:10:32.929 "raid": { 00:10:32.929 "uuid": "6faddf11-01b1-41c9-add2-fa8f648df3e4", 00:10:32.929 "strip_size_kb": 0, 00:10:32.929 "state": "online", 00:10:32.929 "raid_level": "raid1", 00:10:32.929 "superblock": true, 00:10:32.929 "num_base_bdevs": 3, 00:10:32.929 "num_base_bdevs_discovered": 3, 00:10:32.929 "num_base_bdevs_operational": 3, 00:10:32.929 "base_bdevs_list": [ 00:10:32.929 { 00:10:32.929 "name": "pt1", 00:10:32.929 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:32.929 "is_configured": true, 00:10:32.929 "data_offset": 2048, 00:10:32.929 "data_size": 63488 00:10:32.929 }, 00:10:32.929 { 00:10:32.929 "name": "pt2", 00:10:32.929 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:32.929 "is_configured": true, 00:10:32.929 "data_offset": 2048, 00:10:32.929 "data_size": 63488 00:10:32.929 }, 00:10:32.929 { 00:10:32.929 "name": "pt3", 00:10:32.929 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:32.929 "is_configured": true, 00:10:32.929 "data_offset": 2048, 00:10:32.929 "data_size": 63488 00:10:32.929 } 00:10:32.929 ] 00:10:32.929 } 00:10:32.929 } 00:10:32.929 }' 00:10:32.929 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:32.929 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:10:32.929 pt2 00:10:32.929 pt3' 00:10:32.929 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:32.929 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:32.929 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:32.929 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:10:32.929 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:32.929 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:32.929 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.929 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:32.929 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:32.929 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:32.929 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.189 [2024-11-19 15:16:23.374856] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=6faddf11-01b1-41c9-add2-fa8f648df3e4 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 6faddf11-01b1-41c9-add2-fa8f648df3e4 ']' 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.189 [2024-11-19 15:16:23.414570] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:33.189 [2024-11-19 15:16:23.414593] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:33.189 [2024-11-19 15:16:23.414672] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:33.189 [2024-11-19 15:16:23.414746] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:33.189 [2024-11-19 15:16:23.414760] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:33.189 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.450 [2024-11-19 15:16:23.562351] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:10:33.450 [2024-11-19 15:16:23.564581] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:10:33.450 [2024-11-19 15:16:23.564631] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:10:33.450 [2024-11-19 15:16:23.564685] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:10:33.450 [2024-11-19 15:16:23.564737] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:10:33.450 [2024-11-19 15:16:23.564757] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:10:33.450 [2024-11-19 15:16:23.564771] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:33.450 [2024-11-19 15:16:23.564781] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:10:33.450 request: 00:10:33.450 { 00:10:33.450 "name": "raid_bdev1", 00:10:33.450 "raid_level": "raid1", 00:10:33.450 "base_bdevs": [ 00:10:33.450 "malloc1", 00:10:33.450 "malloc2", 00:10:33.450 "malloc3" 00:10:33.450 ], 00:10:33.450 "superblock": false, 00:10:33.450 "method": "bdev_raid_create", 00:10:33.450 "req_id": 1 00:10:33.450 } 00:10:33.450 Got JSON-RPC error response 00:10:33.450 response: 00:10:33.450 { 00:10:33.450 "code": -17, 00:10:33.450 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:10:33.450 } 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.450 [2024-11-19 15:16:23.626222] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:33.450 [2024-11-19 15:16:23.626272] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:33.450 [2024-11-19 15:16:23.626289] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:10:33.450 [2024-11-19 15:16:23.626302] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:33.450 [2024-11-19 15:16:23.628894] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:33.450 [2024-11-19 15:16:23.628952] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:33.450 [2024-11-19 15:16:23.629028] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:10:33.450 [2024-11-19 15:16:23.629071] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:33.450 pt1 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:33.450 "name": "raid_bdev1", 00:10:33.450 "uuid": "6faddf11-01b1-41c9-add2-fa8f648df3e4", 00:10:33.450 "strip_size_kb": 0, 00:10:33.450 "state": "configuring", 00:10:33.450 "raid_level": "raid1", 00:10:33.450 "superblock": true, 00:10:33.450 "num_base_bdevs": 3, 00:10:33.450 "num_base_bdevs_discovered": 1, 00:10:33.450 "num_base_bdevs_operational": 3, 00:10:33.450 "base_bdevs_list": [ 00:10:33.450 { 00:10:33.450 "name": "pt1", 00:10:33.450 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:33.450 "is_configured": true, 00:10:33.450 "data_offset": 2048, 00:10:33.450 "data_size": 63488 00:10:33.450 }, 00:10:33.450 { 00:10:33.450 "name": null, 00:10:33.450 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:33.450 "is_configured": false, 00:10:33.450 "data_offset": 2048, 00:10:33.450 "data_size": 63488 00:10:33.450 }, 00:10:33.450 { 00:10:33.450 "name": null, 00:10:33.450 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:33.450 "is_configured": false, 00:10:33.450 "data_offset": 2048, 00:10:33.450 "data_size": 63488 00:10:33.450 } 00:10:33.450 ] 00:10:33.450 }' 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:33.450 15:16:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.042 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:10:34.042 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:34.042 15:16:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.042 15:16:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.042 [2024-11-19 15:16:24.077457] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:34.042 [2024-11-19 15:16:24.077533] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:34.042 [2024-11-19 15:16:24.077555] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:10:34.042 [2024-11-19 15:16:24.077570] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:34.042 [2024-11-19 15:16:24.078033] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:34.042 [2024-11-19 15:16:24.078055] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:34.042 [2024-11-19 15:16:24.078133] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:34.042 [2024-11-19 15:16:24.078158] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:34.042 pt2 00:10:34.042 15:16:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.042 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:10:34.042 15:16:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.042 15:16:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.042 [2024-11-19 15:16:24.085443] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:10:34.042 15:16:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.042 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:10:34.042 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:34.042 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:34.042 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:34.042 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:34.042 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:34.042 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:34.042 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:34.042 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:34.042 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:34.042 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:34.042 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:34.042 15:16:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.042 15:16:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.042 15:16:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.042 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:34.042 "name": "raid_bdev1", 00:10:34.042 "uuid": "6faddf11-01b1-41c9-add2-fa8f648df3e4", 00:10:34.042 "strip_size_kb": 0, 00:10:34.042 "state": "configuring", 00:10:34.042 "raid_level": "raid1", 00:10:34.042 "superblock": true, 00:10:34.042 "num_base_bdevs": 3, 00:10:34.042 "num_base_bdevs_discovered": 1, 00:10:34.042 "num_base_bdevs_operational": 3, 00:10:34.042 "base_bdevs_list": [ 00:10:34.042 { 00:10:34.042 "name": "pt1", 00:10:34.042 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:34.042 "is_configured": true, 00:10:34.042 "data_offset": 2048, 00:10:34.042 "data_size": 63488 00:10:34.042 }, 00:10:34.042 { 00:10:34.042 "name": null, 00:10:34.042 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:34.042 "is_configured": false, 00:10:34.042 "data_offset": 0, 00:10:34.042 "data_size": 63488 00:10:34.042 }, 00:10:34.042 { 00:10:34.042 "name": null, 00:10:34.042 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:34.042 "is_configured": false, 00:10:34.042 "data_offset": 2048, 00:10:34.043 "data_size": 63488 00:10:34.043 } 00:10:34.043 ] 00:10:34.043 }' 00:10:34.043 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:34.043 15:16:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.303 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:10:34.303 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:34.303 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:34.303 15:16:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.303 15:16:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.303 [2024-11-19 15:16:24.524635] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:34.303 [2024-11-19 15:16:24.524690] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:34.303 [2024-11-19 15:16:24.524710] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:10:34.303 [2024-11-19 15:16:24.524719] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:34.303 [2024-11-19 15:16:24.525215] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:34.303 [2024-11-19 15:16:24.525240] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:34.303 [2024-11-19 15:16:24.525315] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:34.303 [2024-11-19 15:16:24.525341] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:34.303 pt2 00:10:34.303 15:16:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.303 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:34.303 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:34.303 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:34.303 15:16:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.303 15:16:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.303 [2024-11-19 15:16:24.532626] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:34.303 [2024-11-19 15:16:24.532670] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:34.303 [2024-11-19 15:16:24.532692] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:10:34.303 [2024-11-19 15:16:24.532700] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:34.303 [2024-11-19 15:16:24.533061] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:34.303 [2024-11-19 15:16:24.533085] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:34.303 [2024-11-19 15:16:24.533141] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:10:34.303 [2024-11-19 15:16:24.533158] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:34.303 [2024-11-19 15:16:24.533253] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:10:34.303 [2024-11-19 15:16:24.533266] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:34.303 [2024-11-19 15:16:24.533519] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:10:34.303 [2024-11-19 15:16:24.533642] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:10:34.303 [2024-11-19 15:16:24.533657] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:10:34.303 [2024-11-19 15:16:24.533759] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:34.303 pt3 00:10:34.303 15:16:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.303 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:34.303 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:34.303 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:10:34.303 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:34.303 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:34.303 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:34.303 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:34.303 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:34.303 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:34.303 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:34.303 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:34.303 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:34.303 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:34.303 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:34.303 15:16:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.303 15:16:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.303 15:16:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.303 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:34.303 "name": "raid_bdev1", 00:10:34.303 "uuid": "6faddf11-01b1-41c9-add2-fa8f648df3e4", 00:10:34.303 "strip_size_kb": 0, 00:10:34.303 "state": "online", 00:10:34.303 "raid_level": "raid1", 00:10:34.303 "superblock": true, 00:10:34.303 "num_base_bdevs": 3, 00:10:34.303 "num_base_bdevs_discovered": 3, 00:10:34.303 "num_base_bdevs_operational": 3, 00:10:34.303 "base_bdevs_list": [ 00:10:34.303 { 00:10:34.303 "name": "pt1", 00:10:34.303 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:34.303 "is_configured": true, 00:10:34.303 "data_offset": 2048, 00:10:34.303 "data_size": 63488 00:10:34.303 }, 00:10:34.303 { 00:10:34.303 "name": "pt2", 00:10:34.303 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:34.303 "is_configured": true, 00:10:34.303 "data_offset": 2048, 00:10:34.303 "data_size": 63488 00:10:34.303 }, 00:10:34.303 { 00:10:34.303 "name": "pt3", 00:10:34.303 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:34.303 "is_configured": true, 00:10:34.303 "data_offset": 2048, 00:10:34.303 "data_size": 63488 00:10:34.303 } 00:10:34.303 ] 00:10:34.303 }' 00:10:34.303 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:34.303 15:16:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.874 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:10:34.874 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:10:34.874 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:34.874 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:34.874 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:34.874 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:34.874 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:34.874 15:16:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:34.874 15:16:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.874 15:16:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.874 [2024-11-19 15:16:24.976142] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:34.874 15:16:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.874 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:34.874 "name": "raid_bdev1", 00:10:34.874 "aliases": [ 00:10:34.874 "6faddf11-01b1-41c9-add2-fa8f648df3e4" 00:10:34.874 ], 00:10:34.874 "product_name": "Raid Volume", 00:10:34.874 "block_size": 512, 00:10:34.874 "num_blocks": 63488, 00:10:34.874 "uuid": "6faddf11-01b1-41c9-add2-fa8f648df3e4", 00:10:34.874 "assigned_rate_limits": { 00:10:34.874 "rw_ios_per_sec": 0, 00:10:34.874 "rw_mbytes_per_sec": 0, 00:10:34.874 "r_mbytes_per_sec": 0, 00:10:34.874 "w_mbytes_per_sec": 0 00:10:34.874 }, 00:10:34.874 "claimed": false, 00:10:34.874 "zoned": false, 00:10:34.874 "supported_io_types": { 00:10:34.875 "read": true, 00:10:34.875 "write": true, 00:10:34.875 "unmap": false, 00:10:34.875 "flush": false, 00:10:34.875 "reset": true, 00:10:34.875 "nvme_admin": false, 00:10:34.875 "nvme_io": false, 00:10:34.875 "nvme_io_md": false, 00:10:34.875 "write_zeroes": true, 00:10:34.875 "zcopy": false, 00:10:34.875 "get_zone_info": false, 00:10:34.875 "zone_management": false, 00:10:34.875 "zone_append": false, 00:10:34.875 "compare": false, 00:10:34.875 "compare_and_write": false, 00:10:34.875 "abort": false, 00:10:34.875 "seek_hole": false, 00:10:34.875 "seek_data": false, 00:10:34.875 "copy": false, 00:10:34.875 "nvme_iov_md": false 00:10:34.875 }, 00:10:34.875 "memory_domains": [ 00:10:34.875 { 00:10:34.875 "dma_device_id": "system", 00:10:34.875 "dma_device_type": 1 00:10:34.875 }, 00:10:34.875 { 00:10:34.875 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:34.875 "dma_device_type": 2 00:10:34.875 }, 00:10:34.875 { 00:10:34.875 "dma_device_id": "system", 00:10:34.875 "dma_device_type": 1 00:10:34.875 }, 00:10:34.875 { 00:10:34.875 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:34.875 "dma_device_type": 2 00:10:34.875 }, 00:10:34.875 { 00:10:34.875 "dma_device_id": "system", 00:10:34.875 "dma_device_type": 1 00:10:34.875 }, 00:10:34.875 { 00:10:34.875 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:34.875 "dma_device_type": 2 00:10:34.875 } 00:10:34.875 ], 00:10:34.875 "driver_specific": { 00:10:34.875 "raid": { 00:10:34.875 "uuid": "6faddf11-01b1-41c9-add2-fa8f648df3e4", 00:10:34.875 "strip_size_kb": 0, 00:10:34.875 "state": "online", 00:10:34.875 "raid_level": "raid1", 00:10:34.875 "superblock": true, 00:10:34.875 "num_base_bdevs": 3, 00:10:34.875 "num_base_bdevs_discovered": 3, 00:10:34.875 "num_base_bdevs_operational": 3, 00:10:34.875 "base_bdevs_list": [ 00:10:34.875 { 00:10:34.875 "name": "pt1", 00:10:34.875 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:34.875 "is_configured": true, 00:10:34.875 "data_offset": 2048, 00:10:34.875 "data_size": 63488 00:10:34.875 }, 00:10:34.875 { 00:10:34.875 "name": "pt2", 00:10:34.875 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:34.875 "is_configured": true, 00:10:34.875 "data_offset": 2048, 00:10:34.875 "data_size": 63488 00:10:34.875 }, 00:10:34.875 { 00:10:34.875 "name": "pt3", 00:10:34.875 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:34.875 "is_configured": true, 00:10:34.875 "data_offset": 2048, 00:10:34.875 "data_size": 63488 00:10:34.875 } 00:10:34.875 ] 00:10:34.875 } 00:10:34.875 } 00:10:34.875 }' 00:10:34.875 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:34.875 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:10:34.875 pt2 00:10:34.875 pt3' 00:10:34.875 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:34.875 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:34.875 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:34.875 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:10:34.875 15:16:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.875 15:16:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.875 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:34.875 15:16:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.875 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:34.875 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:34.875 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:34.875 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:10:34.875 15:16:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.875 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:34.875 15:16:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.875 15:16:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.135 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:35.135 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:35.135 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:35.136 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:35.136 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:10:35.136 15:16:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.136 15:16:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.136 15:16:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.136 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:35.136 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:35.136 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:35.136 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:10:35.136 15:16:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.136 15:16:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.136 [2024-11-19 15:16:25.271574] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:35.136 15:16:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.136 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 6faddf11-01b1-41c9-add2-fa8f648df3e4 '!=' 6faddf11-01b1-41c9-add2-fa8f648df3e4 ']' 00:10:35.136 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:10:35.136 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:35.136 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:10:35.136 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:10:35.136 15:16:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.136 15:16:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.136 [2024-11-19 15:16:25.315325] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:10:35.136 15:16:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.136 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:10:35.136 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:35.136 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:35.136 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:35.136 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:35.136 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:10:35.136 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:35.136 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:35.136 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:35.136 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:35.136 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:35.136 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:35.136 15:16:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.136 15:16:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.136 15:16:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.136 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:35.136 "name": "raid_bdev1", 00:10:35.136 "uuid": "6faddf11-01b1-41c9-add2-fa8f648df3e4", 00:10:35.136 "strip_size_kb": 0, 00:10:35.136 "state": "online", 00:10:35.136 "raid_level": "raid1", 00:10:35.136 "superblock": true, 00:10:35.136 "num_base_bdevs": 3, 00:10:35.136 "num_base_bdevs_discovered": 2, 00:10:35.136 "num_base_bdevs_operational": 2, 00:10:35.136 "base_bdevs_list": [ 00:10:35.136 { 00:10:35.136 "name": null, 00:10:35.136 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:35.136 "is_configured": false, 00:10:35.136 "data_offset": 0, 00:10:35.136 "data_size": 63488 00:10:35.136 }, 00:10:35.136 { 00:10:35.136 "name": "pt2", 00:10:35.136 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:35.136 "is_configured": true, 00:10:35.136 "data_offset": 2048, 00:10:35.136 "data_size": 63488 00:10:35.136 }, 00:10:35.136 { 00:10:35.136 "name": "pt3", 00:10:35.136 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:35.136 "is_configured": true, 00:10:35.136 "data_offset": 2048, 00:10:35.136 "data_size": 63488 00:10:35.136 } 00:10:35.136 ] 00:10:35.136 }' 00:10:35.136 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:35.136 15:16:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.705 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:35.705 15:16:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.705 15:16:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.705 [2024-11-19 15:16:25.762583] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:35.705 [2024-11-19 15:16:25.762622] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:35.705 [2024-11-19 15:16:25.762694] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:35.705 [2024-11-19 15:16:25.762754] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:35.705 [2024-11-19 15:16:25.762769] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:10:35.705 15:16:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.705 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:35.705 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:10:35.705 15:16:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.705 15:16:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.705 15:16:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.705 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:10:35.705 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:10:35.705 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:10:35.705 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:10:35.705 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:10:35.705 15:16:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.705 15:16:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.705 15:16:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.705 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:10:35.705 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:10:35.705 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:10:35.705 15:16:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.705 15:16:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.705 15:16:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.705 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:10:35.705 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:10:35.705 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:10:35.705 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:10:35.705 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:35.705 15:16:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.705 15:16:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.705 [2024-11-19 15:16:25.846422] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:35.705 [2024-11-19 15:16:25.846474] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:35.705 [2024-11-19 15:16:25.846494] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009980 00:10:35.705 [2024-11-19 15:16:25.846503] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:35.705 [2024-11-19 15:16:25.849159] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:35.705 [2024-11-19 15:16:25.849190] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:35.705 [2024-11-19 15:16:25.849266] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:35.705 [2024-11-19 15:16:25.849301] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:35.705 pt2 00:10:35.705 15:16:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.705 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:10:35.705 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:35.705 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:35.705 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:35.705 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:35.705 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:10:35.705 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:35.706 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:35.706 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:35.706 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:35.706 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:35.706 15:16:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.706 15:16:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.706 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:35.706 15:16:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.706 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:35.706 "name": "raid_bdev1", 00:10:35.706 "uuid": "6faddf11-01b1-41c9-add2-fa8f648df3e4", 00:10:35.706 "strip_size_kb": 0, 00:10:35.706 "state": "configuring", 00:10:35.706 "raid_level": "raid1", 00:10:35.706 "superblock": true, 00:10:35.706 "num_base_bdevs": 3, 00:10:35.706 "num_base_bdevs_discovered": 1, 00:10:35.706 "num_base_bdevs_operational": 2, 00:10:35.706 "base_bdevs_list": [ 00:10:35.706 { 00:10:35.706 "name": null, 00:10:35.706 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:35.706 "is_configured": false, 00:10:35.706 "data_offset": 2048, 00:10:35.706 "data_size": 63488 00:10:35.706 }, 00:10:35.706 { 00:10:35.706 "name": "pt2", 00:10:35.706 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:35.706 "is_configured": true, 00:10:35.706 "data_offset": 2048, 00:10:35.706 "data_size": 63488 00:10:35.706 }, 00:10:35.706 { 00:10:35.706 "name": null, 00:10:35.706 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:35.706 "is_configured": false, 00:10:35.706 "data_offset": 2048, 00:10:35.706 "data_size": 63488 00:10:35.706 } 00:10:35.706 ] 00:10:35.706 }' 00:10:35.706 15:16:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:35.706 15:16:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.966 15:16:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:10:35.966 15:16:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:10:35.966 15:16:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@519 -- # i=2 00:10:35.966 15:16:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:35.966 15:16:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.966 15:16:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.966 [2024-11-19 15:16:26.265763] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:35.966 [2024-11-19 15:16:26.265854] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:35.966 [2024-11-19 15:16:26.265880] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:10:35.966 [2024-11-19 15:16:26.265890] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:35.966 [2024-11-19 15:16:26.266375] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:35.966 [2024-11-19 15:16:26.266392] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:35.966 [2024-11-19 15:16:26.266477] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:10:35.966 [2024-11-19 15:16:26.266508] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:35.966 [2024-11-19 15:16:26.266613] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:10:35.966 [2024-11-19 15:16:26.266621] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:35.966 [2024-11-19 15:16:26.266907] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:10:35.966 [2024-11-19 15:16:26.267082] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:10:35.966 [2024-11-19 15:16:26.267101] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:10:35.966 [2024-11-19 15:16:26.267224] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:35.966 pt3 00:10:35.966 15:16:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.966 15:16:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:10:35.966 15:16:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:35.966 15:16:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:35.966 15:16:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:35.966 15:16:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:35.966 15:16:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:10:35.966 15:16:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:35.966 15:16:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:35.966 15:16:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:35.966 15:16:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:35.966 15:16:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:35.966 15:16:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:35.966 15:16:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.966 15:16:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.966 15:16:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:36.225 15:16:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:36.225 "name": "raid_bdev1", 00:10:36.225 "uuid": "6faddf11-01b1-41c9-add2-fa8f648df3e4", 00:10:36.225 "strip_size_kb": 0, 00:10:36.225 "state": "online", 00:10:36.225 "raid_level": "raid1", 00:10:36.225 "superblock": true, 00:10:36.225 "num_base_bdevs": 3, 00:10:36.225 "num_base_bdevs_discovered": 2, 00:10:36.225 "num_base_bdevs_operational": 2, 00:10:36.225 "base_bdevs_list": [ 00:10:36.225 { 00:10:36.225 "name": null, 00:10:36.225 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:36.225 "is_configured": false, 00:10:36.225 "data_offset": 2048, 00:10:36.225 "data_size": 63488 00:10:36.225 }, 00:10:36.225 { 00:10:36.225 "name": "pt2", 00:10:36.225 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:36.225 "is_configured": true, 00:10:36.225 "data_offset": 2048, 00:10:36.225 "data_size": 63488 00:10:36.225 }, 00:10:36.225 { 00:10:36.225 "name": "pt3", 00:10:36.225 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:36.225 "is_configured": true, 00:10:36.225 "data_offset": 2048, 00:10:36.225 "data_size": 63488 00:10:36.225 } 00:10:36.225 ] 00:10:36.225 }' 00:10:36.225 15:16:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:36.225 15:16:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.485 15:16:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:36.485 15:16:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:36.485 15:16:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.485 [2024-11-19 15:16:26.724955] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:36.485 [2024-11-19 15:16:26.725010] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:36.485 [2024-11-19 15:16:26.725116] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:36.485 [2024-11-19 15:16:26.725202] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:36.485 [2024-11-19 15:16:26.725219] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:10:36.485 15:16:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:36.485 15:16:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:10:36.485 15:16:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:36.485 15:16:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:36.485 15:16:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.485 15:16:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:36.485 15:16:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:10:36.485 15:16:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:10:36.485 15:16:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 3 -gt 2 ']' 00:10:36.485 15:16:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@534 -- # i=2 00:10:36.485 15:16:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt3 00:10:36.485 15:16:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:36.485 15:16:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.485 15:16:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:36.485 15:16:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:36.485 15:16:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:36.485 15:16:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.485 [2024-11-19 15:16:26.776846] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:36.485 [2024-11-19 15:16:26.776920] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:36.485 [2024-11-19 15:16:26.776939] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:10:36.485 [2024-11-19 15:16:26.776951] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:36.485 [2024-11-19 15:16:26.779611] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:36.485 [2024-11-19 15:16:26.779648] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:36.485 [2024-11-19 15:16:26.779777] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:10:36.485 [2024-11-19 15:16:26.779836] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:36.485 [2024-11-19 15:16:26.779983] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:10:36.485 [2024-11-19 15:16:26.780006] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:36.485 [2024-11-19 15:16:26.780029] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:10:36.485 [2024-11-19 15:16:26.780082] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:36.485 pt1 00:10:36.485 15:16:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:36.485 15:16:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 3 -gt 2 ']' 00:10:36.485 15:16:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:10:36.485 15:16:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:36.485 15:16:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:36.485 15:16:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:36.485 15:16:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:36.485 15:16:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:10:36.485 15:16:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:36.485 15:16:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:36.485 15:16:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:36.485 15:16:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:36.485 15:16:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:36.485 15:16:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:36.485 15:16:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:36.485 15:16:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.485 15:16:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:36.744 15:16:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:36.745 "name": "raid_bdev1", 00:10:36.745 "uuid": "6faddf11-01b1-41c9-add2-fa8f648df3e4", 00:10:36.745 "strip_size_kb": 0, 00:10:36.745 "state": "configuring", 00:10:36.745 "raid_level": "raid1", 00:10:36.745 "superblock": true, 00:10:36.745 "num_base_bdevs": 3, 00:10:36.745 "num_base_bdevs_discovered": 1, 00:10:36.745 "num_base_bdevs_operational": 2, 00:10:36.745 "base_bdevs_list": [ 00:10:36.745 { 00:10:36.745 "name": null, 00:10:36.745 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:36.745 "is_configured": false, 00:10:36.745 "data_offset": 2048, 00:10:36.745 "data_size": 63488 00:10:36.745 }, 00:10:36.745 { 00:10:36.745 "name": "pt2", 00:10:36.745 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:36.745 "is_configured": true, 00:10:36.745 "data_offset": 2048, 00:10:36.745 "data_size": 63488 00:10:36.745 }, 00:10:36.745 { 00:10:36.745 "name": null, 00:10:36.745 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:36.745 "is_configured": false, 00:10:36.745 "data_offset": 2048, 00:10:36.745 "data_size": 63488 00:10:36.745 } 00:10:36.745 ] 00:10:36.745 }' 00:10:36.745 15:16:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:36.745 15:16:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.004 15:16:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:10:37.004 15:16:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:10:37.004 15:16:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:37.004 15:16:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.004 15:16:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:37.004 15:16:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:10:37.004 15:16:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:37.004 15:16:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:37.004 15:16:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.004 [2024-11-19 15:16:27.244085] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:37.004 [2024-11-19 15:16:27.244176] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:37.004 [2024-11-19 15:16:27.244197] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:10:37.004 [2024-11-19 15:16:27.244211] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:37.005 [2024-11-19 15:16:27.244699] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:37.005 [2024-11-19 15:16:27.244725] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:37.005 [2024-11-19 15:16:27.244810] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:10:37.005 [2024-11-19 15:16:27.244839] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:37.005 [2024-11-19 15:16:27.244941] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:10:37.005 [2024-11-19 15:16:27.244953] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:37.005 [2024-11-19 15:16:27.245226] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:10:37.005 [2024-11-19 15:16:27.245370] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:10:37.005 [2024-11-19 15:16:27.245379] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:10:37.005 [2024-11-19 15:16:27.245489] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:37.005 pt3 00:10:37.005 15:16:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:37.005 15:16:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:10:37.005 15:16:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:37.005 15:16:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:37.005 15:16:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:37.005 15:16:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:37.005 15:16:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:10:37.005 15:16:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:37.005 15:16:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:37.005 15:16:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:37.005 15:16:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:37.005 15:16:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:37.005 15:16:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:37.005 15:16:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:37.005 15:16:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.005 15:16:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:37.005 15:16:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:37.005 "name": "raid_bdev1", 00:10:37.005 "uuid": "6faddf11-01b1-41c9-add2-fa8f648df3e4", 00:10:37.005 "strip_size_kb": 0, 00:10:37.005 "state": "online", 00:10:37.005 "raid_level": "raid1", 00:10:37.005 "superblock": true, 00:10:37.005 "num_base_bdevs": 3, 00:10:37.005 "num_base_bdevs_discovered": 2, 00:10:37.005 "num_base_bdevs_operational": 2, 00:10:37.005 "base_bdevs_list": [ 00:10:37.005 { 00:10:37.005 "name": null, 00:10:37.005 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:37.005 "is_configured": false, 00:10:37.005 "data_offset": 2048, 00:10:37.005 "data_size": 63488 00:10:37.005 }, 00:10:37.005 { 00:10:37.005 "name": "pt2", 00:10:37.005 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:37.005 "is_configured": true, 00:10:37.005 "data_offset": 2048, 00:10:37.005 "data_size": 63488 00:10:37.005 }, 00:10:37.005 { 00:10:37.005 "name": "pt3", 00:10:37.005 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:37.005 "is_configured": true, 00:10:37.005 "data_offset": 2048, 00:10:37.005 "data_size": 63488 00:10:37.005 } 00:10:37.005 ] 00:10:37.005 }' 00:10:37.005 15:16:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:37.005 15:16:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.575 15:16:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:10:37.575 15:16:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:10:37.575 15:16:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:37.575 15:16:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.575 15:16:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:37.575 15:16:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:10:37.575 15:16:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:37.575 15:16:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:10:37.575 15:16:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:37.575 15:16:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.575 [2024-11-19 15:16:27.720041] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:37.575 15:16:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:37.575 15:16:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' 6faddf11-01b1-41c9-add2-fa8f648df3e4 '!=' 6faddf11-01b1-41c9-add2-fa8f648df3e4 ']' 00:10:37.575 15:16:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 79662 00:10:37.575 15:16:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 79662 ']' 00:10:37.575 15:16:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 79662 00:10:37.575 15:16:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:10:37.575 15:16:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:37.575 15:16:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 79662 00:10:37.575 killing process with pid 79662 00:10:37.575 15:16:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:37.575 15:16:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:37.575 15:16:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 79662' 00:10:37.575 15:16:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 79662 00:10:37.575 15:16:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 79662 00:10:37.575 [2024-11-19 15:16:27.795396] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:37.575 [2024-11-19 15:16:27.795524] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:37.575 [2024-11-19 15:16:27.795612] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:37.575 [2024-11-19 15:16:27.795627] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:10:37.575 [2024-11-19 15:16:27.857869] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:38.145 ************************************ 00:10:38.145 END TEST raid_superblock_test 00:10:38.145 ************************************ 00:10:38.145 15:16:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:10:38.145 00:10:38.145 real 0m6.633s 00:10:38.145 user 0m11.014s 00:10:38.145 sys 0m1.409s 00:10:38.145 15:16:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:38.145 15:16:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.145 15:16:28 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid1 3 read 00:10:38.145 15:16:28 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:10:38.145 15:16:28 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:38.145 15:16:28 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:38.145 ************************************ 00:10:38.145 START TEST raid_read_error_test 00:10:38.145 ************************************ 00:10:38.145 15:16:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid1 3 read 00:10:38.145 15:16:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:10:38.145 15:16:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:10:38.145 15:16:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:10:38.145 15:16:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:10:38.145 15:16:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:38.145 15:16:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:10:38.145 15:16:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:38.145 15:16:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:38.145 15:16:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:10:38.145 15:16:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:38.145 15:16:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:38.145 15:16:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:10:38.145 15:16:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:38.145 15:16:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:38.145 15:16:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:10:38.145 15:16:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:10:38.145 15:16:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:10:38.145 15:16:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:10:38.145 15:16:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:10:38.145 15:16:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:10:38.145 15:16:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:10:38.145 15:16:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:10:38.145 15:16:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:10:38.145 15:16:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:10:38.145 15:16:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.il0CNMfYGJ 00:10:38.145 15:16:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:10:38.145 15:16:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=80097 00:10:38.145 15:16:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 80097 00:10:38.145 15:16:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 80097 ']' 00:10:38.145 15:16:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:38.145 15:16:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:38.145 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:38.145 15:16:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:38.145 15:16:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:38.145 15:16:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.145 [2024-11-19 15:16:28.346142] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:10:38.145 [2024-11-19 15:16:28.346308] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid80097 ] 00:10:38.405 [2024-11-19 15:16:28.485350] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:38.405 [2024-11-19 15:16:28.525987] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:38.405 [2024-11-19 15:16:28.601766] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:38.405 [2024-11-19 15:16:28.601806] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:38.992 15:16:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:38.992 15:16:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:10:38.992 15:16:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:38.992 15:16:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:10:38.992 15:16:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:38.992 15:16:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.992 BaseBdev1_malloc 00:10:38.992 15:16:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:38.992 15:16:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:10:38.992 15:16:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:38.992 15:16:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.992 true 00:10:38.992 15:16:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:38.992 15:16:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:10:38.992 15:16:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:38.992 15:16:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.992 [2024-11-19 15:16:29.223538] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:10:38.992 [2024-11-19 15:16:29.223607] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:38.992 [2024-11-19 15:16:29.223630] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:10:38.992 [2024-11-19 15:16:29.223639] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:38.992 [2024-11-19 15:16:29.226121] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:38.992 [2024-11-19 15:16:29.226155] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:38.992 BaseBdev1 00:10:38.992 15:16:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:38.992 15:16:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:38.992 15:16:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:10:38.992 15:16:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:38.992 15:16:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.992 BaseBdev2_malloc 00:10:38.992 15:16:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:38.992 15:16:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:10:38.992 15:16:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:38.992 15:16:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.992 true 00:10:38.992 15:16:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:38.992 15:16:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:10:38.992 15:16:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:38.992 15:16:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.992 [2024-11-19 15:16:29.270158] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:10:38.992 [2024-11-19 15:16:29.270214] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:38.992 [2024-11-19 15:16:29.270233] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:10:38.992 [2024-11-19 15:16:29.270251] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:38.992 [2024-11-19 15:16:29.272646] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:38.992 [2024-11-19 15:16:29.272684] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:10:38.992 BaseBdev2 00:10:38.992 15:16:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:38.992 15:16:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:38.992 15:16:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:10:38.992 15:16:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:38.992 15:16:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.992 BaseBdev3_malloc 00:10:38.992 15:16:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:38.992 15:16:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:10:38.992 15:16:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:38.992 15:16:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.992 true 00:10:38.992 15:16:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:38.992 15:16:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:10:38.992 15:16:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:38.992 15:16:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.992 [2024-11-19 15:16:29.316774] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:10:38.992 [2024-11-19 15:16:29.316830] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:38.992 [2024-11-19 15:16:29.316851] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:10:38.992 [2024-11-19 15:16:29.316860] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:38.992 [2024-11-19 15:16:29.319239] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:38.992 [2024-11-19 15:16:29.319269] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:10:38.992 BaseBdev3 00:10:38.992 15:16:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:38.992 15:16:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:10:38.992 15:16:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:38.992 15:16:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.992 [2024-11-19 15:16:29.328844] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:39.254 [2024-11-19 15:16:29.331010] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:39.254 [2024-11-19 15:16:29.331092] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:39.254 [2024-11-19 15:16:29.331274] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:10:39.254 [2024-11-19 15:16:29.331294] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:39.254 [2024-11-19 15:16:29.331563] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002bb0 00:10:39.254 [2024-11-19 15:16:29.331750] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:10:39.255 [2024-11-19 15:16:29.331768] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:10:39.255 [2024-11-19 15:16:29.331924] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:39.255 15:16:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:39.255 15:16:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:10:39.255 15:16:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:39.255 15:16:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:39.255 15:16:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:39.255 15:16:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:39.255 15:16:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:39.255 15:16:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:39.255 15:16:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:39.255 15:16:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:39.255 15:16:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:39.255 15:16:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:39.255 15:16:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:39.255 15:16:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:39.255 15:16:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.255 15:16:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:39.255 15:16:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:39.255 "name": "raid_bdev1", 00:10:39.255 "uuid": "1802db61-5a7e-43ed-82a4-7c1c92e3a5c7", 00:10:39.255 "strip_size_kb": 0, 00:10:39.255 "state": "online", 00:10:39.255 "raid_level": "raid1", 00:10:39.255 "superblock": true, 00:10:39.255 "num_base_bdevs": 3, 00:10:39.255 "num_base_bdevs_discovered": 3, 00:10:39.255 "num_base_bdevs_operational": 3, 00:10:39.255 "base_bdevs_list": [ 00:10:39.255 { 00:10:39.255 "name": "BaseBdev1", 00:10:39.255 "uuid": "4f9f592f-ce80-5f9a-95bf-b442ce7f90b5", 00:10:39.255 "is_configured": true, 00:10:39.255 "data_offset": 2048, 00:10:39.255 "data_size": 63488 00:10:39.255 }, 00:10:39.255 { 00:10:39.255 "name": "BaseBdev2", 00:10:39.255 "uuid": "20b708fa-de3e-548e-bd06-d1544c1c85e6", 00:10:39.255 "is_configured": true, 00:10:39.255 "data_offset": 2048, 00:10:39.255 "data_size": 63488 00:10:39.255 }, 00:10:39.255 { 00:10:39.255 "name": "BaseBdev3", 00:10:39.255 "uuid": "c8eaa96e-049f-52d8-b7ef-10d53a1f5ea7", 00:10:39.255 "is_configured": true, 00:10:39.255 "data_offset": 2048, 00:10:39.255 "data_size": 63488 00:10:39.255 } 00:10:39.255 ] 00:10:39.255 }' 00:10:39.255 15:16:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:39.255 15:16:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.514 15:16:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:10:39.514 15:16:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:10:39.772 [2024-11-19 15:16:29.892659] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002d50 00:10:40.711 15:16:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:10:40.711 15:16:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:40.711 15:16:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.711 15:16:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:40.711 15:16:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:10:40.711 15:16:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:10:40.711 15:16:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ read = \w\r\i\t\e ]] 00:10:40.711 15:16:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:10:40.711 15:16:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:10:40.711 15:16:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:40.711 15:16:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:40.711 15:16:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:40.711 15:16:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:40.711 15:16:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:40.711 15:16:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:40.711 15:16:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:40.711 15:16:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:40.711 15:16:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:40.711 15:16:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:40.711 15:16:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:40.711 15:16:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:40.711 15:16:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.711 15:16:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:40.711 15:16:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:40.711 "name": "raid_bdev1", 00:10:40.711 "uuid": "1802db61-5a7e-43ed-82a4-7c1c92e3a5c7", 00:10:40.711 "strip_size_kb": 0, 00:10:40.711 "state": "online", 00:10:40.711 "raid_level": "raid1", 00:10:40.711 "superblock": true, 00:10:40.711 "num_base_bdevs": 3, 00:10:40.711 "num_base_bdevs_discovered": 3, 00:10:40.711 "num_base_bdevs_operational": 3, 00:10:40.711 "base_bdevs_list": [ 00:10:40.711 { 00:10:40.711 "name": "BaseBdev1", 00:10:40.711 "uuid": "4f9f592f-ce80-5f9a-95bf-b442ce7f90b5", 00:10:40.711 "is_configured": true, 00:10:40.711 "data_offset": 2048, 00:10:40.711 "data_size": 63488 00:10:40.711 }, 00:10:40.711 { 00:10:40.711 "name": "BaseBdev2", 00:10:40.711 "uuid": "20b708fa-de3e-548e-bd06-d1544c1c85e6", 00:10:40.711 "is_configured": true, 00:10:40.711 "data_offset": 2048, 00:10:40.711 "data_size": 63488 00:10:40.711 }, 00:10:40.711 { 00:10:40.711 "name": "BaseBdev3", 00:10:40.711 "uuid": "c8eaa96e-049f-52d8-b7ef-10d53a1f5ea7", 00:10:40.711 "is_configured": true, 00:10:40.711 "data_offset": 2048, 00:10:40.711 "data_size": 63488 00:10:40.711 } 00:10:40.711 ] 00:10:40.711 }' 00:10:40.711 15:16:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:40.711 15:16:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.970 15:16:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:40.970 15:16:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:40.970 15:16:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.970 [2024-11-19 15:16:31.261963] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:40.970 [2024-11-19 15:16:31.262027] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:40.970 [2024-11-19 15:16:31.264541] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:40.970 [2024-11-19 15:16:31.264608] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:40.970 [2024-11-19 15:16:31.264717] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:40.970 [2024-11-19 15:16:31.264733] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:10:40.970 { 00:10:40.970 "results": [ 00:10:40.970 { 00:10:40.970 "job": "raid_bdev1", 00:10:40.970 "core_mask": "0x1", 00:10:40.970 "workload": "randrw", 00:10:40.970 "percentage": 50, 00:10:40.970 "status": "finished", 00:10:40.970 "queue_depth": 1, 00:10:40.970 "io_size": 131072, 00:10:40.970 "runtime": 1.36975, 00:10:40.970 "iops": 10985.216280343127, 00:10:40.970 "mibps": 1373.152035042891, 00:10:40.970 "io_failed": 0, 00:10:40.970 "io_timeout": 0, 00:10:40.970 "avg_latency_us": 88.53243406467594, 00:10:40.970 "min_latency_us": 22.022707423580787, 00:10:40.970 "max_latency_us": 1402.2986899563318 00:10:40.970 } 00:10:40.970 ], 00:10:40.970 "core_count": 1 00:10:40.970 } 00:10:40.970 15:16:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:40.970 15:16:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 80097 00:10:40.970 15:16:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 80097 ']' 00:10:40.970 15:16:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 80097 00:10:40.970 15:16:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:10:40.970 15:16:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:40.970 15:16:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 80097 00:10:41.229 15:16:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:41.229 15:16:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:41.229 killing process with pid 80097 00:10:41.229 15:16:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 80097' 00:10:41.229 15:16:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 80097 00:10:41.229 [2024-11-19 15:16:31.309292] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:41.229 15:16:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 80097 00:10:41.229 [2024-11-19 15:16:31.358388] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:41.489 15:16:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:10:41.489 15:16:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.il0CNMfYGJ 00:10:41.489 15:16:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:10:41.489 15:16:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:10:41.489 15:16:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:10:41.489 15:16:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:41.489 15:16:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:10:41.489 15:16:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:10:41.489 00:10:41.489 real 0m3.430s 00:10:41.489 user 0m4.266s 00:10:41.489 sys 0m0.608s 00:10:41.489 15:16:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:41.489 15:16:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.489 ************************************ 00:10:41.489 END TEST raid_read_error_test 00:10:41.489 ************************************ 00:10:41.489 15:16:31 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid1 3 write 00:10:41.489 15:16:31 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:10:41.489 15:16:31 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:41.489 15:16:31 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:41.489 ************************************ 00:10:41.489 START TEST raid_write_error_test 00:10:41.489 ************************************ 00:10:41.489 15:16:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid1 3 write 00:10:41.489 15:16:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:10:41.489 15:16:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:10:41.489 15:16:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:10:41.489 15:16:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:10:41.489 15:16:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:41.489 15:16:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:10:41.489 15:16:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:41.489 15:16:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:41.489 15:16:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:10:41.489 15:16:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:41.489 15:16:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:41.489 15:16:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:10:41.489 15:16:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:41.489 15:16:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:41.489 15:16:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:10:41.489 15:16:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:10:41.489 15:16:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:10:41.489 15:16:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:10:41.489 15:16:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:10:41.489 15:16:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:10:41.489 15:16:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:10:41.489 15:16:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:10:41.489 15:16:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:10:41.489 15:16:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:10:41.489 15:16:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.GzNBO363lY 00:10:41.489 15:16:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=80226 00:10:41.489 15:16:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 80226 00:10:41.489 15:16:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:10:41.489 15:16:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 80226 ']' 00:10:41.489 15:16:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:41.489 15:16:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:41.489 15:16:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:41.489 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:41.489 15:16:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:41.489 15:16:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.749 [2024-11-19 15:16:31.846946] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:10:41.749 [2024-11-19 15:16:31.847091] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid80226 ] 00:10:41.749 [2024-11-19 15:16:32.000991] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:41.749 [2024-11-19 15:16:32.042093] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:42.009 [2024-11-19 15:16:32.118573] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:42.009 [2024-11-19 15:16:32.118613] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:42.590 15:16:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:42.590 15:16:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:10:42.590 15:16:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:42.590 15:16:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:10:42.590 15:16:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.590 15:16:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.590 BaseBdev1_malloc 00:10:42.590 15:16:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.590 15:16:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:10:42.590 15:16:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.590 15:16:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.590 true 00:10:42.590 15:16:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.590 15:16:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:10:42.590 15:16:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.590 15:16:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.590 [2024-11-19 15:16:32.716578] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:10:42.590 [2024-11-19 15:16:32.716643] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:42.590 [2024-11-19 15:16:32.716671] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:10:42.590 [2024-11-19 15:16:32.716679] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:42.590 [2024-11-19 15:16:32.719106] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:42.590 [2024-11-19 15:16:32.719137] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:42.590 BaseBdev1 00:10:42.590 15:16:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.590 15:16:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:42.590 15:16:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:10:42.590 15:16:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.590 15:16:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.590 BaseBdev2_malloc 00:10:42.590 15:16:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.590 15:16:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:10:42.590 15:16:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.590 15:16:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.590 true 00:10:42.590 15:16:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.590 15:16:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:10:42.590 15:16:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.590 15:16:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.590 [2024-11-19 15:16:32.763089] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:10:42.590 [2024-11-19 15:16:32.763142] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:42.590 [2024-11-19 15:16:32.763162] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:10:42.590 [2024-11-19 15:16:32.763181] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:42.590 [2024-11-19 15:16:32.765695] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:42.590 [2024-11-19 15:16:32.765734] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:10:42.590 BaseBdev2 00:10:42.590 15:16:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.590 15:16:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:42.590 15:16:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:10:42.590 15:16:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.590 15:16:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.591 BaseBdev3_malloc 00:10:42.591 15:16:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.591 15:16:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:10:42.591 15:16:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.591 15:16:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.591 true 00:10:42.591 15:16:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.591 15:16:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:10:42.591 15:16:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.591 15:16:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.591 [2024-11-19 15:16:32.809595] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:10:42.591 [2024-11-19 15:16:32.809644] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:42.591 [2024-11-19 15:16:32.809664] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:10:42.591 [2024-11-19 15:16:32.809672] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:42.591 [2024-11-19 15:16:32.812004] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:42.591 [2024-11-19 15:16:32.812036] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:10:42.591 BaseBdev3 00:10:42.591 15:16:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.591 15:16:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:10:42.591 15:16:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.591 15:16:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.591 [2024-11-19 15:16:32.821651] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:42.591 [2024-11-19 15:16:32.823695] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:42.591 [2024-11-19 15:16:32.823803] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:42.591 [2024-11-19 15:16:32.824004] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:10:42.591 [2024-11-19 15:16:32.824024] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:42.591 [2024-11-19 15:16:32.824271] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002bb0 00:10:42.591 [2024-11-19 15:16:32.824451] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:10:42.591 [2024-11-19 15:16:32.824467] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:10:42.591 [2024-11-19 15:16:32.824595] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:42.591 15:16:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.591 15:16:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:10:42.591 15:16:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:42.591 15:16:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:42.591 15:16:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:42.591 15:16:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:42.591 15:16:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:42.591 15:16:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:42.591 15:16:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:42.591 15:16:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:42.591 15:16:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:42.591 15:16:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:42.591 15:16:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:42.591 15:16:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.591 15:16:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.591 15:16:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.591 15:16:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:42.591 "name": "raid_bdev1", 00:10:42.591 "uuid": "7374531a-e5fa-4327-ab3a-9488e9b08351", 00:10:42.591 "strip_size_kb": 0, 00:10:42.591 "state": "online", 00:10:42.591 "raid_level": "raid1", 00:10:42.591 "superblock": true, 00:10:42.591 "num_base_bdevs": 3, 00:10:42.591 "num_base_bdevs_discovered": 3, 00:10:42.591 "num_base_bdevs_operational": 3, 00:10:42.591 "base_bdevs_list": [ 00:10:42.591 { 00:10:42.591 "name": "BaseBdev1", 00:10:42.591 "uuid": "c50e35f1-c5a9-53b2-b65f-32c92fa6a131", 00:10:42.591 "is_configured": true, 00:10:42.591 "data_offset": 2048, 00:10:42.591 "data_size": 63488 00:10:42.591 }, 00:10:42.591 { 00:10:42.591 "name": "BaseBdev2", 00:10:42.591 "uuid": "de00085c-d6f4-52be-8be6-9cea1a8a6366", 00:10:42.591 "is_configured": true, 00:10:42.591 "data_offset": 2048, 00:10:42.591 "data_size": 63488 00:10:42.591 }, 00:10:42.591 { 00:10:42.591 "name": "BaseBdev3", 00:10:42.591 "uuid": "95d68fa3-a3f1-5e60-86ee-fa9a021f1efb", 00:10:42.591 "is_configured": true, 00:10:42.591 "data_offset": 2048, 00:10:42.591 "data_size": 63488 00:10:42.591 } 00:10:42.591 ] 00:10:42.591 }' 00:10:42.591 15:16:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:42.591 15:16:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:43.178 15:16:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:10:43.178 15:16:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:10:43.178 [2024-11-19 15:16:33.333407] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002d50 00:10:44.119 15:16:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:10:44.119 15:16:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:44.119 15:16:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.119 [2024-11-19 15:16:34.248059] bdev_raid.c:2276:_raid_bdev_fail_base_bdev: *NOTICE*: Failing base bdev in slot 0 ('BaseBdev1') of raid bdev 'raid_bdev1' 00:10:44.119 [2024-11-19 15:16:34.248148] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:44.119 [2024-11-19 15:16:34.248395] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000002d50 00:10:44.119 15:16:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:44.119 15:16:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:10:44.119 15:16:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:10:44.119 15:16:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ write = \w\r\i\t\e ]] 00:10:44.119 15:16:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=2 00:10:44.119 15:16:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:10:44.119 15:16:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:44.119 15:16:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:44.119 15:16:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:44.119 15:16:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:44.119 15:16:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:10:44.119 15:16:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:44.119 15:16:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:44.119 15:16:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:44.119 15:16:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:44.119 15:16:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:44.119 15:16:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:44.119 15:16:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:44.119 15:16:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.119 15:16:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:44.119 15:16:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:44.119 "name": "raid_bdev1", 00:10:44.119 "uuid": "7374531a-e5fa-4327-ab3a-9488e9b08351", 00:10:44.119 "strip_size_kb": 0, 00:10:44.119 "state": "online", 00:10:44.119 "raid_level": "raid1", 00:10:44.119 "superblock": true, 00:10:44.119 "num_base_bdevs": 3, 00:10:44.119 "num_base_bdevs_discovered": 2, 00:10:44.119 "num_base_bdevs_operational": 2, 00:10:44.119 "base_bdevs_list": [ 00:10:44.119 { 00:10:44.119 "name": null, 00:10:44.119 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:44.119 "is_configured": false, 00:10:44.119 "data_offset": 0, 00:10:44.119 "data_size": 63488 00:10:44.119 }, 00:10:44.119 { 00:10:44.119 "name": "BaseBdev2", 00:10:44.119 "uuid": "de00085c-d6f4-52be-8be6-9cea1a8a6366", 00:10:44.119 "is_configured": true, 00:10:44.119 "data_offset": 2048, 00:10:44.119 "data_size": 63488 00:10:44.119 }, 00:10:44.119 { 00:10:44.119 "name": "BaseBdev3", 00:10:44.119 "uuid": "95d68fa3-a3f1-5e60-86ee-fa9a021f1efb", 00:10:44.119 "is_configured": true, 00:10:44.119 "data_offset": 2048, 00:10:44.119 "data_size": 63488 00:10:44.119 } 00:10:44.119 ] 00:10:44.119 }' 00:10:44.119 15:16:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:44.119 15:16:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.380 15:16:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:44.380 15:16:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:44.380 15:16:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.380 [2024-11-19 15:16:34.666561] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:44.380 [2024-11-19 15:16:34.666610] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:44.380 [2024-11-19 15:16:34.669170] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:44.380 [2024-11-19 15:16:34.669223] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:44.380 [2024-11-19 15:16:34.669325] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:44.380 [2024-11-19 15:16:34.669336] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:10:44.380 { 00:10:44.380 "results": [ 00:10:44.380 { 00:10:44.380 "job": "raid_bdev1", 00:10:44.380 "core_mask": "0x1", 00:10:44.380 "workload": "randrw", 00:10:44.380 "percentage": 50, 00:10:44.380 "status": "finished", 00:10:44.380 "queue_depth": 1, 00:10:44.380 "io_size": 131072, 00:10:44.380 "runtime": 1.333617, 00:10:44.380 "iops": 12594.320558301222, 00:10:44.380 "mibps": 1574.2900697876528, 00:10:44.380 "io_failed": 0, 00:10:44.380 "io_timeout": 0, 00:10:44.380 "avg_latency_us": 76.87172262890624, 00:10:44.380 "min_latency_us": 21.799126637554586, 00:10:44.380 "max_latency_us": 1316.4436681222708 00:10:44.380 } 00:10:44.380 ], 00:10:44.380 "core_count": 1 00:10:44.380 } 00:10:44.380 15:16:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:44.380 15:16:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 80226 00:10:44.380 15:16:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 80226 ']' 00:10:44.380 15:16:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 80226 00:10:44.380 15:16:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:10:44.380 15:16:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:44.380 15:16:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 80226 00:10:44.380 15:16:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:44.380 15:16:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:44.380 killing process with pid 80226 00:10:44.380 15:16:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 80226' 00:10:44.380 15:16:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 80226 00:10:44.380 [2024-11-19 15:16:34.710698] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:44.380 15:16:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 80226 00:10:44.640 [2024-11-19 15:16:34.758529] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:44.899 15:16:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.GzNBO363lY 00:10:44.899 15:16:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:10:44.900 15:16:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:10:44.900 15:16:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:10:44.900 15:16:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:10:44.900 15:16:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:44.900 15:16:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:10:44.900 15:16:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:10:44.900 00:10:44.900 real 0m3.336s 00:10:44.900 user 0m4.103s 00:10:44.900 sys 0m0.588s 00:10:44.900 15:16:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:44.900 15:16:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.900 ************************************ 00:10:44.900 END TEST raid_write_error_test 00:10:44.900 ************************************ 00:10:44.900 15:16:35 bdev_raid -- bdev/bdev_raid.sh@966 -- # for n in {2..4} 00:10:44.900 15:16:35 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:10:44.900 15:16:35 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid0 4 false 00:10:44.900 15:16:35 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:10:44.900 15:16:35 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:44.900 15:16:35 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:44.900 ************************************ 00:10:44.900 START TEST raid_state_function_test 00:10:44.900 ************************************ 00:10:44.900 15:16:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid0 4 false 00:10:44.900 15:16:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:10:44.900 15:16:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:10:44.900 15:16:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:10:44.900 15:16:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:10:44.900 15:16:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:10:44.900 15:16:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:44.900 15:16:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:10:44.900 15:16:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:44.900 15:16:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:44.900 15:16:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:10:44.900 15:16:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:44.900 15:16:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:44.900 15:16:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:10:44.900 15:16:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:44.900 15:16:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:44.900 15:16:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:10:44.900 15:16:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:44.900 15:16:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:44.900 15:16:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:44.900 15:16:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:10:44.900 15:16:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:10:44.900 15:16:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:10:44.900 15:16:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:10:44.900 15:16:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:10:44.900 15:16:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:10:44.900 15:16:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:10:44.900 15:16:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:10:44.900 15:16:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:10:44.900 15:16:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:10:44.900 15:16:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=80356 00:10:44.900 15:16:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:10:44.900 15:16:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 80356' 00:10:44.900 Process raid pid: 80356 00:10:44.900 15:16:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 80356 00:10:44.900 15:16:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 80356 ']' 00:10:44.900 15:16:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:44.900 15:16:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:44.900 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:44.900 15:16:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:44.900 15:16:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:44.900 15:16:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:45.160 [2024-11-19 15:16:35.251502] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:10:45.160 [2024-11-19 15:16:35.251621] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:45.160 [2024-11-19 15:16:35.409254] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:45.160 [2024-11-19 15:16:35.448958] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:45.420 [2024-11-19 15:16:35.525039] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:45.420 [2024-11-19 15:16:35.525079] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:45.990 15:16:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:45.990 15:16:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:10:45.990 15:16:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:45.990 15:16:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:45.990 15:16:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:45.990 [2024-11-19 15:16:36.088074] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:45.990 [2024-11-19 15:16:36.088138] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:45.990 [2024-11-19 15:16:36.088149] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:45.990 [2024-11-19 15:16:36.088159] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:45.990 [2024-11-19 15:16:36.088165] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:45.990 [2024-11-19 15:16:36.088178] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:45.990 [2024-11-19 15:16:36.088184] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:45.990 [2024-11-19 15:16:36.088193] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:45.990 15:16:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:45.990 15:16:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:45.990 15:16:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:45.990 15:16:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:45.990 15:16:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:45.990 15:16:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:45.990 15:16:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:45.991 15:16:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:45.991 15:16:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:45.991 15:16:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:45.991 15:16:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:45.991 15:16:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:45.991 15:16:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:45.991 15:16:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:45.991 15:16:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:45.991 15:16:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:45.991 15:16:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:45.991 "name": "Existed_Raid", 00:10:45.991 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:45.991 "strip_size_kb": 64, 00:10:45.991 "state": "configuring", 00:10:45.991 "raid_level": "raid0", 00:10:45.991 "superblock": false, 00:10:45.991 "num_base_bdevs": 4, 00:10:45.991 "num_base_bdevs_discovered": 0, 00:10:45.991 "num_base_bdevs_operational": 4, 00:10:45.991 "base_bdevs_list": [ 00:10:45.991 { 00:10:45.991 "name": "BaseBdev1", 00:10:45.991 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:45.991 "is_configured": false, 00:10:45.991 "data_offset": 0, 00:10:45.991 "data_size": 0 00:10:45.991 }, 00:10:45.991 { 00:10:45.991 "name": "BaseBdev2", 00:10:45.991 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:45.991 "is_configured": false, 00:10:45.991 "data_offset": 0, 00:10:45.991 "data_size": 0 00:10:45.991 }, 00:10:45.991 { 00:10:45.991 "name": "BaseBdev3", 00:10:45.991 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:45.991 "is_configured": false, 00:10:45.991 "data_offset": 0, 00:10:45.991 "data_size": 0 00:10:45.991 }, 00:10:45.991 { 00:10:45.991 "name": "BaseBdev4", 00:10:45.991 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:45.991 "is_configured": false, 00:10:45.991 "data_offset": 0, 00:10:45.991 "data_size": 0 00:10:45.991 } 00:10:45.991 ] 00:10:45.991 }' 00:10:45.991 15:16:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:45.991 15:16:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.252 15:16:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:46.252 15:16:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:46.252 15:16:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.252 [2024-11-19 15:16:36.523342] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:46.252 [2024-11-19 15:16:36.523398] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:10:46.252 15:16:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:46.252 15:16:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:46.252 15:16:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:46.252 15:16:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.252 [2024-11-19 15:16:36.535316] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:46.252 [2024-11-19 15:16:36.535362] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:46.252 [2024-11-19 15:16:36.535371] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:46.252 [2024-11-19 15:16:36.535381] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:46.252 [2024-11-19 15:16:36.535387] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:46.252 [2024-11-19 15:16:36.535397] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:46.252 [2024-11-19 15:16:36.535403] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:46.252 [2024-11-19 15:16:36.535412] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:46.252 15:16:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:46.252 15:16:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:46.252 15:16:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:46.252 15:16:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.252 [2024-11-19 15:16:36.562334] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:46.252 BaseBdev1 00:10:46.252 15:16:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:46.252 15:16:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:10:46.252 15:16:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:10:46.252 15:16:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:46.252 15:16:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:46.252 15:16:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:46.252 15:16:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:46.252 15:16:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:46.252 15:16:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:46.252 15:16:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.252 15:16:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:46.252 15:16:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:46.252 15:16:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:46.252 15:16:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.512 [ 00:10:46.512 { 00:10:46.512 "name": "BaseBdev1", 00:10:46.512 "aliases": [ 00:10:46.512 "88fd03b6-9212-40ac-8af3-bffe459dae35" 00:10:46.512 ], 00:10:46.512 "product_name": "Malloc disk", 00:10:46.512 "block_size": 512, 00:10:46.512 "num_blocks": 65536, 00:10:46.512 "uuid": "88fd03b6-9212-40ac-8af3-bffe459dae35", 00:10:46.512 "assigned_rate_limits": { 00:10:46.512 "rw_ios_per_sec": 0, 00:10:46.512 "rw_mbytes_per_sec": 0, 00:10:46.512 "r_mbytes_per_sec": 0, 00:10:46.512 "w_mbytes_per_sec": 0 00:10:46.512 }, 00:10:46.512 "claimed": true, 00:10:46.512 "claim_type": "exclusive_write", 00:10:46.512 "zoned": false, 00:10:46.512 "supported_io_types": { 00:10:46.512 "read": true, 00:10:46.512 "write": true, 00:10:46.512 "unmap": true, 00:10:46.512 "flush": true, 00:10:46.512 "reset": true, 00:10:46.512 "nvme_admin": false, 00:10:46.512 "nvme_io": false, 00:10:46.512 "nvme_io_md": false, 00:10:46.512 "write_zeroes": true, 00:10:46.512 "zcopy": true, 00:10:46.512 "get_zone_info": false, 00:10:46.512 "zone_management": false, 00:10:46.512 "zone_append": false, 00:10:46.512 "compare": false, 00:10:46.512 "compare_and_write": false, 00:10:46.512 "abort": true, 00:10:46.512 "seek_hole": false, 00:10:46.512 "seek_data": false, 00:10:46.512 "copy": true, 00:10:46.512 "nvme_iov_md": false 00:10:46.512 }, 00:10:46.512 "memory_domains": [ 00:10:46.512 { 00:10:46.512 "dma_device_id": "system", 00:10:46.512 "dma_device_type": 1 00:10:46.512 }, 00:10:46.512 { 00:10:46.512 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:46.512 "dma_device_type": 2 00:10:46.512 } 00:10:46.512 ], 00:10:46.512 "driver_specific": {} 00:10:46.512 } 00:10:46.512 ] 00:10:46.512 15:16:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:46.512 15:16:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:46.512 15:16:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:46.512 15:16:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:46.512 15:16:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:46.512 15:16:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:46.512 15:16:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:46.512 15:16:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:46.512 15:16:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:46.512 15:16:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:46.512 15:16:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:46.512 15:16:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:46.512 15:16:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:46.512 15:16:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:46.512 15:16:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:46.512 15:16:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.512 15:16:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:46.512 15:16:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:46.512 "name": "Existed_Raid", 00:10:46.512 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:46.512 "strip_size_kb": 64, 00:10:46.512 "state": "configuring", 00:10:46.512 "raid_level": "raid0", 00:10:46.512 "superblock": false, 00:10:46.512 "num_base_bdevs": 4, 00:10:46.512 "num_base_bdevs_discovered": 1, 00:10:46.512 "num_base_bdevs_operational": 4, 00:10:46.512 "base_bdevs_list": [ 00:10:46.512 { 00:10:46.512 "name": "BaseBdev1", 00:10:46.512 "uuid": "88fd03b6-9212-40ac-8af3-bffe459dae35", 00:10:46.512 "is_configured": true, 00:10:46.512 "data_offset": 0, 00:10:46.512 "data_size": 65536 00:10:46.512 }, 00:10:46.512 { 00:10:46.512 "name": "BaseBdev2", 00:10:46.512 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:46.512 "is_configured": false, 00:10:46.512 "data_offset": 0, 00:10:46.512 "data_size": 0 00:10:46.512 }, 00:10:46.512 { 00:10:46.512 "name": "BaseBdev3", 00:10:46.512 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:46.512 "is_configured": false, 00:10:46.512 "data_offset": 0, 00:10:46.512 "data_size": 0 00:10:46.512 }, 00:10:46.512 { 00:10:46.512 "name": "BaseBdev4", 00:10:46.512 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:46.512 "is_configured": false, 00:10:46.512 "data_offset": 0, 00:10:46.512 "data_size": 0 00:10:46.512 } 00:10:46.512 ] 00:10:46.512 }' 00:10:46.512 15:16:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:46.512 15:16:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.773 15:16:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:46.773 15:16:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:46.773 15:16:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.773 [2024-11-19 15:16:36.997651] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:46.773 [2024-11-19 15:16:36.997724] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:10:46.773 15:16:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:46.773 15:16:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:46.773 15:16:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:46.773 15:16:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.773 [2024-11-19 15:16:37.005670] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:46.773 [2024-11-19 15:16:37.007842] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:46.773 [2024-11-19 15:16:37.007880] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:46.773 [2024-11-19 15:16:37.007890] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:46.773 [2024-11-19 15:16:37.007899] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:46.773 [2024-11-19 15:16:37.007905] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:46.773 [2024-11-19 15:16:37.007913] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:46.773 15:16:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:46.773 15:16:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:10:46.773 15:16:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:46.773 15:16:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:46.773 15:16:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:46.773 15:16:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:46.773 15:16:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:46.773 15:16:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:46.773 15:16:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:46.773 15:16:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:46.773 15:16:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:46.773 15:16:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:46.773 15:16:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:46.773 15:16:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:46.773 15:16:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:46.773 15:16:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:46.773 15:16:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.773 15:16:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:46.773 15:16:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:46.773 "name": "Existed_Raid", 00:10:46.773 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:46.773 "strip_size_kb": 64, 00:10:46.773 "state": "configuring", 00:10:46.773 "raid_level": "raid0", 00:10:46.773 "superblock": false, 00:10:46.773 "num_base_bdevs": 4, 00:10:46.773 "num_base_bdevs_discovered": 1, 00:10:46.773 "num_base_bdevs_operational": 4, 00:10:46.773 "base_bdevs_list": [ 00:10:46.773 { 00:10:46.773 "name": "BaseBdev1", 00:10:46.773 "uuid": "88fd03b6-9212-40ac-8af3-bffe459dae35", 00:10:46.773 "is_configured": true, 00:10:46.773 "data_offset": 0, 00:10:46.773 "data_size": 65536 00:10:46.773 }, 00:10:46.773 { 00:10:46.773 "name": "BaseBdev2", 00:10:46.773 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:46.773 "is_configured": false, 00:10:46.773 "data_offset": 0, 00:10:46.773 "data_size": 0 00:10:46.773 }, 00:10:46.773 { 00:10:46.773 "name": "BaseBdev3", 00:10:46.773 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:46.773 "is_configured": false, 00:10:46.773 "data_offset": 0, 00:10:46.773 "data_size": 0 00:10:46.773 }, 00:10:46.773 { 00:10:46.773 "name": "BaseBdev4", 00:10:46.773 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:46.773 "is_configured": false, 00:10:46.773 "data_offset": 0, 00:10:46.773 "data_size": 0 00:10:46.773 } 00:10:46.773 ] 00:10:46.773 }' 00:10:46.773 15:16:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:46.773 15:16:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:47.342 15:16:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:47.342 15:16:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:47.342 15:16:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:47.342 [2024-11-19 15:16:37.445747] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:47.342 BaseBdev2 00:10:47.342 15:16:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:47.342 15:16:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:10:47.342 15:16:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:10:47.342 15:16:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:47.342 15:16:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:47.342 15:16:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:47.342 15:16:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:47.342 15:16:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:47.342 15:16:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:47.342 15:16:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:47.342 15:16:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:47.342 15:16:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:47.342 15:16:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:47.342 15:16:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:47.342 [ 00:10:47.342 { 00:10:47.342 "name": "BaseBdev2", 00:10:47.342 "aliases": [ 00:10:47.342 "8c5b044f-6f56-47a6-853f-faac83736767" 00:10:47.342 ], 00:10:47.342 "product_name": "Malloc disk", 00:10:47.342 "block_size": 512, 00:10:47.342 "num_blocks": 65536, 00:10:47.342 "uuid": "8c5b044f-6f56-47a6-853f-faac83736767", 00:10:47.342 "assigned_rate_limits": { 00:10:47.342 "rw_ios_per_sec": 0, 00:10:47.342 "rw_mbytes_per_sec": 0, 00:10:47.342 "r_mbytes_per_sec": 0, 00:10:47.342 "w_mbytes_per_sec": 0 00:10:47.342 }, 00:10:47.342 "claimed": true, 00:10:47.342 "claim_type": "exclusive_write", 00:10:47.342 "zoned": false, 00:10:47.342 "supported_io_types": { 00:10:47.342 "read": true, 00:10:47.342 "write": true, 00:10:47.342 "unmap": true, 00:10:47.342 "flush": true, 00:10:47.342 "reset": true, 00:10:47.342 "nvme_admin": false, 00:10:47.342 "nvme_io": false, 00:10:47.342 "nvme_io_md": false, 00:10:47.342 "write_zeroes": true, 00:10:47.342 "zcopy": true, 00:10:47.342 "get_zone_info": false, 00:10:47.342 "zone_management": false, 00:10:47.342 "zone_append": false, 00:10:47.342 "compare": false, 00:10:47.342 "compare_and_write": false, 00:10:47.342 "abort": true, 00:10:47.342 "seek_hole": false, 00:10:47.342 "seek_data": false, 00:10:47.342 "copy": true, 00:10:47.342 "nvme_iov_md": false 00:10:47.342 }, 00:10:47.342 "memory_domains": [ 00:10:47.342 { 00:10:47.342 "dma_device_id": "system", 00:10:47.342 "dma_device_type": 1 00:10:47.342 }, 00:10:47.342 { 00:10:47.342 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:47.342 "dma_device_type": 2 00:10:47.342 } 00:10:47.342 ], 00:10:47.342 "driver_specific": {} 00:10:47.342 } 00:10:47.342 ] 00:10:47.342 15:16:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:47.342 15:16:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:47.342 15:16:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:47.343 15:16:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:47.343 15:16:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:47.343 15:16:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:47.343 15:16:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:47.343 15:16:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:47.343 15:16:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:47.343 15:16:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:47.343 15:16:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:47.343 15:16:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:47.343 15:16:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:47.343 15:16:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:47.343 15:16:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:47.343 15:16:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:47.343 15:16:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:47.343 15:16:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:47.343 15:16:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:47.343 15:16:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:47.343 "name": "Existed_Raid", 00:10:47.343 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:47.343 "strip_size_kb": 64, 00:10:47.343 "state": "configuring", 00:10:47.343 "raid_level": "raid0", 00:10:47.343 "superblock": false, 00:10:47.343 "num_base_bdevs": 4, 00:10:47.343 "num_base_bdevs_discovered": 2, 00:10:47.343 "num_base_bdevs_operational": 4, 00:10:47.343 "base_bdevs_list": [ 00:10:47.343 { 00:10:47.343 "name": "BaseBdev1", 00:10:47.343 "uuid": "88fd03b6-9212-40ac-8af3-bffe459dae35", 00:10:47.343 "is_configured": true, 00:10:47.343 "data_offset": 0, 00:10:47.343 "data_size": 65536 00:10:47.343 }, 00:10:47.343 { 00:10:47.343 "name": "BaseBdev2", 00:10:47.343 "uuid": "8c5b044f-6f56-47a6-853f-faac83736767", 00:10:47.343 "is_configured": true, 00:10:47.343 "data_offset": 0, 00:10:47.343 "data_size": 65536 00:10:47.343 }, 00:10:47.343 { 00:10:47.343 "name": "BaseBdev3", 00:10:47.343 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:47.343 "is_configured": false, 00:10:47.343 "data_offset": 0, 00:10:47.343 "data_size": 0 00:10:47.343 }, 00:10:47.343 { 00:10:47.343 "name": "BaseBdev4", 00:10:47.343 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:47.343 "is_configured": false, 00:10:47.343 "data_offset": 0, 00:10:47.343 "data_size": 0 00:10:47.343 } 00:10:47.343 ] 00:10:47.343 }' 00:10:47.343 15:16:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:47.343 15:16:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:47.912 15:16:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:47.912 15:16:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:47.912 15:16:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:47.912 [2024-11-19 15:16:37.972534] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:47.912 BaseBdev3 00:10:47.912 15:16:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:47.912 15:16:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:10:47.912 15:16:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:10:47.912 15:16:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:47.912 15:16:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:47.912 15:16:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:47.912 15:16:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:47.912 15:16:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:47.912 15:16:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:47.912 15:16:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:47.912 15:16:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:47.912 15:16:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:47.912 15:16:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:47.912 15:16:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:47.912 [ 00:10:47.913 { 00:10:47.913 "name": "BaseBdev3", 00:10:47.913 "aliases": [ 00:10:47.913 "054cae68-bc3d-463a-a5c5-0e2f555ace69" 00:10:47.913 ], 00:10:47.913 "product_name": "Malloc disk", 00:10:47.913 "block_size": 512, 00:10:47.913 "num_blocks": 65536, 00:10:47.913 "uuid": "054cae68-bc3d-463a-a5c5-0e2f555ace69", 00:10:47.913 "assigned_rate_limits": { 00:10:47.913 "rw_ios_per_sec": 0, 00:10:47.913 "rw_mbytes_per_sec": 0, 00:10:47.913 "r_mbytes_per_sec": 0, 00:10:47.913 "w_mbytes_per_sec": 0 00:10:47.913 }, 00:10:47.913 "claimed": true, 00:10:47.913 "claim_type": "exclusive_write", 00:10:47.913 "zoned": false, 00:10:47.913 "supported_io_types": { 00:10:47.913 "read": true, 00:10:47.913 "write": true, 00:10:47.913 "unmap": true, 00:10:47.913 "flush": true, 00:10:47.913 "reset": true, 00:10:47.913 "nvme_admin": false, 00:10:47.913 "nvme_io": false, 00:10:47.913 "nvme_io_md": false, 00:10:47.913 "write_zeroes": true, 00:10:47.913 "zcopy": true, 00:10:47.913 "get_zone_info": false, 00:10:47.913 "zone_management": false, 00:10:47.913 "zone_append": false, 00:10:47.913 "compare": false, 00:10:47.913 "compare_and_write": false, 00:10:47.913 "abort": true, 00:10:47.913 "seek_hole": false, 00:10:47.913 "seek_data": false, 00:10:47.913 "copy": true, 00:10:47.913 "nvme_iov_md": false 00:10:47.913 }, 00:10:47.913 "memory_domains": [ 00:10:47.913 { 00:10:47.913 "dma_device_id": "system", 00:10:47.913 "dma_device_type": 1 00:10:47.913 }, 00:10:47.913 { 00:10:47.913 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:47.913 "dma_device_type": 2 00:10:47.913 } 00:10:47.913 ], 00:10:47.913 "driver_specific": {} 00:10:47.913 } 00:10:47.913 ] 00:10:47.913 15:16:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:47.913 15:16:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:47.913 15:16:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:47.913 15:16:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:47.913 15:16:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:47.913 15:16:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:47.913 15:16:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:47.913 15:16:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:47.913 15:16:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:47.913 15:16:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:47.913 15:16:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:47.913 15:16:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:47.913 15:16:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:47.913 15:16:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:47.913 15:16:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:47.913 15:16:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:47.913 15:16:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:47.913 15:16:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:47.913 15:16:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:47.913 15:16:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:47.913 "name": "Existed_Raid", 00:10:47.913 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:47.913 "strip_size_kb": 64, 00:10:47.913 "state": "configuring", 00:10:47.913 "raid_level": "raid0", 00:10:47.913 "superblock": false, 00:10:47.913 "num_base_bdevs": 4, 00:10:47.913 "num_base_bdevs_discovered": 3, 00:10:47.913 "num_base_bdevs_operational": 4, 00:10:47.913 "base_bdevs_list": [ 00:10:47.913 { 00:10:47.913 "name": "BaseBdev1", 00:10:47.913 "uuid": "88fd03b6-9212-40ac-8af3-bffe459dae35", 00:10:47.913 "is_configured": true, 00:10:47.913 "data_offset": 0, 00:10:47.913 "data_size": 65536 00:10:47.913 }, 00:10:47.913 { 00:10:47.913 "name": "BaseBdev2", 00:10:47.913 "uuid": "8c5b044f-6f56-47a6-853f-faac83736767", 00:10:47.913 "is_configured": true, 00:10:47.913 "data_offset": 0, 00:10:47.913 "data_size": 65536 00:10:47.913 }, 00:10:47.913 { 00:10:47.913 "name": "BaseBdev3", 00:10:47.913 "uuid": "054cae68-bc3d-463a-a5c5-0e2f555ace69", 00:10:47.913 "is_configured": true, 00:10:47.913 "data_offset": 0, 00:10:47.913 "data_size": 65536 00:10:47.913 }, 00:10:47.913 { 00:10:47.913 "name": "BaseBdev4", 00:10:47.913 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:47.913 "is_configured": false, 00:10:47.913 "data_offset": 0, 00:10:47.913 "data_size": 0 00:10:47.913 } 00:10:47.913 ] 00:10:47.913 }' 00:10:47.913 15:16:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:47.913 15:16:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.173 15:16:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:48.173 15:16:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.173 15:16:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.173 [2024-11-19 15:16:38.444693] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:48.173 [2024-11-19 15:16:38.444753] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:10:48.173 [2024-11-19 15:16:38.444764] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:10:48.173 [2024-11-19 15:16:38.445118] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:10:48.173 [2024-11-19 15:16:38.445281] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:10:48.173 [2024-11-19 15:16:38.445304] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:10:48.173 [2024-11-19 15:16:38.445550] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:48.173 BaseBdev4 00:10:48.173 15:16:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.173 15:16:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:10:48.173 15:16:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:10:48.173 15:16:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:48.173 15:16:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:48.173 15:16:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:48.173 15:16:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:48.173 15:16:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:48.173 15:16:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.173 15:16:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.173 15:16:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.173 15:16:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:48.173 15:16:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.173 15:16:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.173 [ 00:10:48.173 { 00:10:48.173 "name": "BaseBdev4", 00:10:48.173 "aliases": [ 00:10:48.173 "06496a63-e93b-4a9a-82ff-8a9d4fbb7a19" 00:10:48.173 ], 00:10:48.173 "product_name": "Malloc disk", 00:10:48.173 "block_size": 512, 00:10:48.173 "num_blocks": 65536, 00:10:48.173 "uuid": "06496a63-e93b-4a9a-82ff-8a9d4fbb7a19", 00:10:48.173 "assigned_rate_limits": { 00:10:48.173 "rw_ios_per_sec": 0, 00:10:48.173 "rw_mbytes_per_sec": 0, 00:10:48.173 "r_mbytes_per_sec": 0, 00:10:48.173 "w_mbytes_per_sec": 0 00:10:48.173 }, 00:10:48.173 "claimed": true, 00:10:48.173 "claim_type": "exclusive_write", 00:10:48.173 "zoned": false, 00:10:48.173 "supported_io_types": { 00:10:48.173 "read": true, 00:10:48.173 "write": true, 00:10:48.173 "unmap": true, 00:10:48.173 "flush": true, 00:10:48.173 "reset": true, 00:10:48.173 "nvme_admin": false, 00:10:48.173 "nvme_io": false, 00:10:48.173 "nvme_io_md": false, 00:10:48.173 "write_zeroes": true, 00:10:48.173 "zcopy": true, 00:10:48.173 "get_zone_info": false, 00:10:48.173 "zone_management": false, 00:10:48.173 "zone_append": false, 00:10:48.173 "compare": false, 00:10:48.173 "compare_and_write": false, 00:10:48.173 "abort": true, 00:10:48.173 "seek_hole": false, 00:10:48.173 "seek_data": false, 00:10:48.173 "copy": true, 00:10:48.173 "nvme_iov_md": false 00:10:48.173 }, 00:10:48.173 "memory_domains": [ 00:10:48.173 { 00:10:48.173 "dma_device_id": "system", 00:10:48.173 "dma_device_type": 1 00:10:48.173 }, 00:10:48.173 { 00:10:48.173 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:48.173 "dma_device_type": 2 00:10:48.173 } 00:10:48.173 ], 00:10:48.173 "driver_specific": {} 00:10:48.173 } 00:10:48.173 ] 00:10:48.173 15:16:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.173 15:16:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:48.173 15:16:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:48.173 15:16:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:48.173 15:16:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:10:48.173 15:16:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:48.173 15:16:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:48.173 15:16:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:48.173 15:16:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:48.174 15:16:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:48.174 15:16:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:48.174 15:16:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:48.174 15:16:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:48.174 15:16:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:48.174 15:16:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:48.174 15:16:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.174 15:16:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:48.174 15:16:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.432 15:16:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.432 15:16:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:48.432 "name": "Existed_Raid", 00:10:48.432 "uuid": "42a4c8fa-7d8e-45e3-bf95-72bc617aadf3", 00:10:48.432 "strip_size_kb": 64, 00:10:48.432 "state": "online", 00:10:48.432 "raid_level": "raid0", 00:10:48.432 "superblock": false, 00:10:48.432 "num_base_bdevs": 4, 00:10:48.432 "num_base_bdevs_discovered": 4, 00:10:48.432 "num_base_bdevs_operational": 4, 00:10:48.432 "base_bdevs_list": [ 00:10:48.432 { 00:10:48.432 "name": "BaseBdev1", 00:10:48.432 "uuid": "88fd03b6-9212-40ac-8af3-bffe459dae35", 00:10:48.432 "is_configured": true, 00:10:48.432 "data_offset": 0, 00:10:48.432 "data_size": 65536 00:10:48.432 }, 00:10:48.432 { 00:10:48.432 "name": "BaseBdev2", 00:10:48.432 "uuid": "8c5b044f-6f56-47a6-853f-faac83736767", 00:10:48.432 "is_configured": true, 00:10:48.432 "data_offset": 0, 00:10:48.432 "data_size": 65536 00:10:48.432 }, 00:10:48.432 { 00:10:48.432 "name": "BaseBdev3", 00:10:48.432 "uuid": "054cae68-bc3d-463a-a5c5-0e2f555ace69", 00:10:48.432 "is_configured": true, 00:10:48.432 "data_offset": 0, 00:10:48.432 "data_size": 65536 00:10:48.432 }, 00:10:48.432 { 00:10:48.432 "name": "BaseBdev4", 00:10:48.432 "uuid": "06496a63-e93b-4a9a-82ff-8a9d4fbb7a19", 00:10:48.432 "is_configured": true, 00:10:48.432 "data_offset": 0, 00:10:48.432 "data_size": 65536 00:10:48.432 } 00:10:48.432 ] 00:10:48.432 }' 00:10:48.432 15:16:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:48.432 15:16:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.692 15:16:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:10:48.692 15:16:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:48.692 15:16:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:48.692 15:16:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:48.692 15:16:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:48.692 15:16:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:48.692 15:16:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:48.692 15:16:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:48.692 15:16:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.692 15:16:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.692 [2024-11-19 15:16:38.944275] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:48.692 15:16:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.692 15:16:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:48.692 "name": "Existed_Raid", 00:10:48.692 "aliases": [ 00:10:48.692 "42a4c8fa-7d8e-45e3-bf95-72bc617aadf3" 00:10:48.692 ], 00:10:48.692 "product_name": "Raid Volume", 00:10:48.692 "block_size": 512, 00:10:48.692 "num_blocks": 262144, 00:10:48.692 "uuid": "42a4c8fa-7d8e-45e3-bf95-72bc617aadf3", 00:10:48.692 "assigned_rate_limits": { 00:10:48.692 "rw_ios_per_sec": 0, 00:10:48.692 "rw_mbytes_per_sec": 0, 00:10:48.692 "r_mbytes_per_sec": 0, 00:10:48.692 "w_mbytes_per_sec": 0 00:10:48.692 }, 00:10:48.692 "claimed": false, 00:10:48.692 "zoned": false, 00:10:48.692 "supported_io_types": { 00:10:48.692 "read": true, 00:10:48.692 "write": true, 00:10:48.692 "unmap": true, 00:10:48.692 "flush": true, 00:10:48.692 "reset": true, 00:10:48.692 "nvme_admin": false, 00:10:48.692 "nvme_io": false, 00:10:48.692 "nvme_io_md": false, 00:10:48.692 "write_zeroes": true, 00:10:48.692 "zcopy": false, 00:10:48.692 "get_zone_info": false, 00:10:48.692 "zone_management": false, 00:10:48.692 "zone_append": false, 00:10:48.692 "compare": false, 00:10:48.692 "compare_and_write": false, 00:10:48.692 "abort": false, 00:10:48.692 "seek_hole": false, 00:10:48.692 "seek_data": false, 00:10:48.692 "copy": false, 00:10:48.692 "nvme_iov_md": false 00:10:48.692 }, 00:10:48.692 "memory_domains": [ 00:10:48.692 { 00:10:48.692 "dma_device_id": "system", 00:10:48.692 "dma_device_type": 1 00:10:48.692 }, 00:10:48.692 { 00:10:48.692 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:48.692 "dma_device_type": 2 00:10:48.692 }, 00:10:48.692 { 00:10:48.692 "dma_device_id": "system", 00:10:48.692 "dma_device_type": 1 00:10:48.692 }, 00:10:48.692 { 00:10:48.692 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:48.692 "dma_device_type": 2 00:10:48.692 }, 00:10:48.692 { 00:10:48.692 "dma_device_id": "system", 00:10:48.692 "dma_device_type": 1 00:10:48.692 }, 00:10:48.692 { 00:10:48.692 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:48.692 "dma_device_type": 2 00:10:48.692 }, 00:10:48.692 { 00:10:48.692 "dma_device_id": "system", 00:10:48.692 "dma_device_type": 1 00:10:48.692 }, 00:10:48.692 { 00:10:48.692 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:48.692 "dma_device_type": 2 00:10:48.692 } 00:10:48.692 ], 00:10:48.692 "driver_specific": { 00:10:48.692 "raid": { 00:10:48.692 "uuid": "42a4c8fa-7d8e-45e3-bf95-72bc617aadf3", 00:10:48.692 "strip_size_kb": 64, 00:10:48.692 "state": "online", 00:10:48.692 "raid_level": "raid0", 00:10:48.692 "superblock": false, 00:10:48.692 "num_base_bdevs": 4, 00:10:48.692 "num_base_bdevs_discovered": 4, 00:10:48.692 "num_base_bdevs_operational": 4, 00:10:48.692 "base_bdevs_list": [ 00:10:48.692 { 00:10:48.692 "name": "BaseBdev1", 00:10:48.692 "uuid": "88fd03b6-9212-40ac-8af3-bffe459dae35", 00:10:48.692 "is_configured": true, 00:10:48.692 "data_offset": 0, 00:10:48.692 "data_size": 65536 00:10:48.692 }, 00:10:48.692 { 00:10:48.692 "name": "BaseBdev2", 00:10:48.692 "uuid": "8c5b044f-6f56-47a6-853f-faac83736767", 00:10:48.692 "is_configured": true, 00:10:48.692 "data_offset": 0, 00:10:48.692 "data_size": 65536 00:10:48.692 }, 00:10:48.692 { 00:10:48.692 "name": "BaseBdev3", 00:10:48.692 "uuid": "054cae68-bc3d-463a-a5c5-0e2f555ace69", 00:10:48.692 "is_configured": true, 00:10:48.692 "data_offset": 0, 00:10:48.692 "data_size": 65536 00:10:48.692 }, 00:10:48.692 { 00:10:48.692 "name": "BaseBdev4", 00:10:48.692 "uuid": "06496a63-e93b-4a9a-82ff-8a9d4fbb7a19", 00:10:48.692 "is_configured": true, 00:10:48.692 "data_offset": 0, 00:10:48.692 "data_size": 65536 00:10:48.692 } 00:10:48.692 ] 00:10:48.692 } 00:10:48.692 } 00:10:48.692 }' 00:10:48.692 15:16:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:48.692 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:10:48.692 BaseBdev2 00:10:48.692 BaseBdev3 00:10:48.692 BaseBdev4' 00:10:48.692 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:48.959 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:48.959 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:48.959 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:48.959 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:10:48.959 15:16:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.959 15:16:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.959 15:16:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.959 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:48.959 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:48.959 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:48.959 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:48.959 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:48.959 15:16:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.959 15:16:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.959 15:16:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.959 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:48.959 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:48.959 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:48.959 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:48.959 15:16:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.959 15:16:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.959 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:48.959 15:16:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.959 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:48.959 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:48.959 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:48.960 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:48.960 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:48.960 15:16:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.960 15:16:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.960 15:16:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.960 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:48.960 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:48.960 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:48.960 15:16:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.960 15:16:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.960 [2024-11-19 15:16:39.263472] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:48.960 [2024-11-19 15:16:39.263517] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:48.960 [2024-11-19 15:16:39.263577] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:48.960 15:16:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.960 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:10:48.960 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:10:48.960 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:48.960 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:48.960 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:10:48.960 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 3 00:10:48.960 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:48.960 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:10:48.960 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:48.960 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:48.960 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:48.960 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:48.960 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:48.960 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:48.960 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:48.960 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:48.960 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:48.960 15:16:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.960 15:16:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.221 15:16:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:49.222 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:49.222 "name": "Existed_Raid", 00:10:49.222 "uuid": "42a4c8fa-7d8e-45e3-bf95-72bc617aadf3", 00:10:49.222 "strip_size_kb": 64, 00:10:49.222 "state": "offline", 00:10:49.222 "raid_level": "raid0", 00:10:49.222 "superblock": false, 00:10:49.222 "num_base_bdevs": 4, 00:10:49.222 "num_base_bdevs_discovered": 3, 00:10:49.222 "num_base_bdevs_operational": 3, 00:10:49.222 "base_bdevs_list": [ 00:10:49.222 { 00:10:49.222 "name": null, 00:10:49.222 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:49.222 "is_configured": false, 00:10:49.222 "data_offset": 0, 00:10:49.222 "data_size": 65536 00:10:49.222 }, 00:10:49.222 { 00:10:49.222 "name": "BaseBdev2", 00:10:49.222 "uuid": "8c5b044f-6f56-47a6-853f-faac83736767", 00:10:49.222 "is_configured": true, 00:10:49.222 "data_offset": 0, 00:10:49.222 "data_size": 65536 00:10:49.222 }, 00:10:49.222 { 00:10:49.222 "name": "BaseBdev3", 00:10:49.222 "uuid": "054cae68-bc3d-463a-a5c5-0e2f555ace69", 00:10:49.222 "is_configured": true, 00:10:49.222 "data_offset": 0, 00:10:49.222 "data_size": 65536 00:10:49.222 }, 00:10:49.222 { 00:10:49.222 "name": "BaseBdev4", 00:10:49.222 "uuid": "06496a63-e93b-4a9a-82ff-8a9d4fbb7a19", 00:10:49.222 "is_configured": true, 00:10:49.222 "data_offset": 0, 00:10:49.222 "data_size": 65536 00:10:49.222 } 00:10:49.222 ] 00:10:49.222 }' 00:10:49.222 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:49.222 15:16:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.482 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:10:49.482 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:49.482 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:49.482 15:16:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:49.482 15:16:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.482 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:49.482 15:16:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:49.482 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:49.482 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:49.482 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:10:49.482 15:16:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:49.482 15:16:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.482 [2024-11-19 15:16:39.786712] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:49.482 15:16:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:49.482 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:49.482 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:49.482 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:49.482 15:16:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:49.482 15:16:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.482 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:49.742 15:16:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:49.742 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:49.742 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:49.742 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:10:49.742 15:16:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:49.742 15:16:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.742 [2024-11-19 15:16:39.863155] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:49.742 15:16:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:49.742 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:49.742 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:49.742 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:49.742 15:16:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:49.742 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:49.742 15:16:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.742 15:16:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:49.742 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:49.742 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:49.742 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:10:49.742 15:16:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:49.742 15:16:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.742 [2024-11-19 15:16:39.939746] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:10:49.742 [2024-11-19 15:16:39.939804] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:10:49.742 15:16:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:49.742 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:49.742 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:49.742 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:49.742 15:16:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:10:49.742 15:16:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:49.742 15:16:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.742 15:16:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:49.742 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:10:49.742 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:10:49.742 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:10:49.742 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:10:49.742 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:49.742 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:49.742 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:49.742 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.742 BaseBdev2 00:10:49.742 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:49.742 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:10:49.743 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:10:49.743 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:49.743 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:49.743 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:49.743 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:49.743 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:49.743 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:49.743 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.743 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:49.743 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:49.743 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:49.743 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.743 [ 00:10:49.743 { 00:10:49.743 "name": "BaseBdev2", 00:10:49.743 "aliases": [ 00:10:49.743 "11d501d0-1e66-41a3-95e7-cf12d4eb94ab" 00:10:49.743 ], 00:10:49.743 "product_name": "Malloc disk", 00:10:49.743 "block_size": 512, 00:10:49.743 "num_blocks": 65536, 00:10:49.743 "uuid": "11d501d0-1e66-41a3-95e7-cf12d4eb94ab", 00:10:49.743 "assigned_rate_limits": { 00:10:49.743 "rw_ios_per_sec": 0, 00:10:49.743 "rw_mbytes_per_sec": 0, 00:10:49.743 "r_mbytes_per_sec": 0, 00:10:49.743 "w_mbytes_per_sec": 0 00:10:49.743 }, 00:10:49.743 "claimed": false, 00:10:49.743 "zoned": false, 00:10:49.743 "supported_io_types": { 00:10:49.743 "read": true, 00:10:49.743 "write": true, 00:10:49.743 "unmap": true, 00:10:49.743 "flush": true, 00:10:49.743 "reset": true, 00:10:49.743 "nvme_admin": false, 00:10:49.743 "nvme_io": false, 00:10:49.743 "nvme_io_md": false, 00:10:49.743 "write_zeroes": true, 00:10:49.743 "zcopy": true, 00:10:49.743 "get_zone_info": false, 00:10:49.743 "zone_management": false, 00:10:49.743 "zone_append": false, 00:10:49.743 "compare": false, 00:10:49.743 "compare_and_write": false, 00:10:49.743 "abort": true, 00:10:49.743 "seek_hole": false, 00:10:49.743 "seek_data": false, 00:10:49.743 "copy": true, 00:10:49.743 "nvme_iov_md": false 00:10:49.743 }, 00:10:49.743 "memory_domains": [ 00:10:49.743 { 00:10:49.743 "dma_device_id": "system", 00:10:49.743 "dma_device_type": 1 00:10:49.743 }, 00:10:49.743 { 00:10:49.743 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:49.743 "dma_device_type": 2 00:10:49.743 } 00:10:49.743 ], 00:10:49.743 "driver_specific": {} 00:10:49.743 } 00:10:49.743 ] 00:10:49.743 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:49.743 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:49.743 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:49.743 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:49.743 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:49.743 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:49.743 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.003 BaseBdev3 00:10:50.003 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:50.003 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:10:50.003 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:10:50.003 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:50.003 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:50.003 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:50.003 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:50.003 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:50.003 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:50.003 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.003 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:50.003 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:50.003 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:50.003 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.003 [ 00:10:50.003 { 00:10:50.003 "name": "BaseBdev3", 00:10:50.003 "aliases": [ 00:10:50.003 "0d02744e-2130-42a1-b37f-0cab8b3471dc" 00:10:50.003 ], 00:10:50.003 "product_name": "Malloc disk", 00:10:50.003 "block_size": 512, 00:10:50.003 "num_blocks": 65536, 00:10:50.003 "uuid": "0d02744e-2130-42a1-b37f-0cab8b3471dc", 00:10:50.003 "assigned_rate_limits": { 00:10:50.003 "rw_ios_per_sec": 0, 00:10:50.003 "rw_mbytes_per_sec": 0, 00:10:50.003 "r_mbytes_per_sec": 0, 00:10:50.003 "w_mbytes_per_sec": 0 00:10:50.003 }, 00:10:50.003 "claimed": false, 00:10:50.003 "zoned": false, 00:10:50.003 "supported_io_types": { 00:10:50.003 "read": true, 00:10:50.003 "write": true, 00:10:50.003 "unmap": true, 00:10:50.003 "flush": true, 00:10:50.003 "reset": true, 00:10:50.003 "nvme_admin": false, 00:10:50.003 "nvme_io": false, 00:10:50.003 "nvme_io_md": false, 00:10:50.003 "write_zeroes": true, 00:10:50.003 "zcopy": true, 00:10:50.003 "get_zone_info": false, 00:10:50.003 "zone_management": false, 00:10:50.003 "zone_append": false, 00:10:50.003 "compare": false, 00:10:50.003 "compare_and_write": false, 00:10:50.003 "abort": true, 00:10:50.003 "seek_hole": false, 00:10:50.003 "seek_data": false, 00:10:50.003 "copy": true, 00:10:50.003 "nvme_iov_md": false 00:10:50.003 }, 00:10:50.003 "memory_domains": [ 00:10:50.003 { 00:10:50.003 "dma_device_id": "system", 00:10:50.003 "dma_device_type": 1 00:10:50.003 }, 00:10:50.003 { 00:10:50.003 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:50.003 "dma_device_type": 2 00:10:50.003 } 00:10:50.003 ], 00:10:50.003 "driver_specific": {} 00:10:50.003 } 00:10:50.003 ] 00:10:50.003 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:50.003 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:50.003 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:50.003 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:50.003 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:50.003 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:50.003 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.003 BaseBdev4 00:10:50.003 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:50.003 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:10:50.003 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:10:50.003 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:50.003 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:50.003 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:50.003 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:50.003 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:50.003 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:50.003 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.003 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:50.003 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:50.003 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:50.003 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.003 [ 00:10:50.003 { 00:10:50.003 "name": "BaseBdev4", 00:10:50.003 "aliases": [ 00:10:50.003 "a5690c56-f60f-4d9e-8387-0fdcf8017f70" 00:10:50.003 ], 00:10:50.003 "product_name": "Malloc disk", 00:10:50.003 "block_size": 512, 00:10:50.003 "num_blocks": 65536, 00:10:50.003 "uuid": "a5690c56-f60f-4d9e-8387-0fdcf8017f70", 00:10:50.003 "assigned_rate_limits": { 00:10:50.003 "rw_ios_per_sec": 0, 00:10:50.003 "rw_mbytes_per_sec": 0, 00:10:50.003 "r_mbytes_per_sec": 0, 00:10:50.003 "w_mbytes_per_sec": 0 00:10:50.003 }, 00:10:50.003 "claimed": false, 00:10:50.003 "zoned": false, 00:10:50.003 "supported_io_types": { 00:10:50.003 "read": true, 00:10:50.003 "write": true, 00:10:50.003 "unmap": true, 00:10:50.003 "flush": true, 00:10:50.003 "reset": true, 00:10:50.003 "nvme_admin": false, 00:10:50.003 "nvme_io": false, 00:10:50.003 "nvme_io_md": false, 00:10:50.003 "write_zeroes": true, 00:10:50.003 "zcopy": true, 00:10:50.003 "get_zone_info": false, 00:10:50.003 "zone_management": false, 00:10:50.003 "zone_append": false, 00:10:50.003 "compare": false, 00:10:50.003 "compare_and_write": false, 00:10:50.003 "abort": true, 00:10:50.003 "seek_hole": false, 00:10:50.004 "seek_data": false, 00:10:50.004 "copy": true, 00:10:50.004 "nvme_iov_md": false 00:10:50.004 }, 00:10:50.004 "memory_domains": [ 00:10:50.004 { 00:10:50.004 "dma_device_id": "system", 00:10:50.004 "dma_device_type": 1 00:10:50.004 }, 00:10:50.004 { 00:10:50.004 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:50.004 "dma_device_type": 2 00:10:50.004 } 00:10:50.004 ], 00:10:50.004 "driver_specific": {} 00:10:50.004 } 00:10:50.004 ] 00:10:50.004 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:50.004 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:50.004 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:50.004 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:50.004 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:50.004 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:50.004 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.004 [2024-11-19 15:16:40.197729] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:50.004 [2024-11-19 15:16:40.197778] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:50.004 [2024-11-19 15:16:40.197818] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:50.004 [2024-11-19 15:16:40.199932] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:50.004 [2024-11-19 15:16:40.199994] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:50.004 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:50.004 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:50.004 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:50.004 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:50.004 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:50.004 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:50.004 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:50.004 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:50.004 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:50.004 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:50.004 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:50.004 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:50.004 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:50.004 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:50.004 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.004 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:50.004 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:50.004 "name": "Existed_Raid", 00:10:50.004 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:50.004 "strip_size_kb": 64, 00:10:50.004 "state": "configuring", 00:10:50.004 "raid_level": "raid0", 00:10:50.004 "superblock": false, 00:10:50.004 "num_base_bdevs": 4, 00:10:50.004 "num_base_bdevs_discovered": 3, 00:10:50.004 "num_base_bdevs_operational": 4, 00:10:50.004 "base_bdevs_list": [ 00:10:50.004 { 00:10:50.004 "name": "BaseBdev1", 00:10:50.004 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:50.004 "is_configured": false, 00:10:50.004 "data_offset": 0, 00:10:50.004 "data_size": 0 00:10:50.004 }, 00:10:50.004 { 00:10:50.004 "name": "BaseBdev2", 00:10:50.004 "uuid": "11d501d0-1e66-41a3-95e7-cf12d4eb94ab", 00:10:50.004 "is_configured": true, 00:10:50.004 "data_offset": 0, 00:10:50.004 "data_size": 65536 00:10:50.004 }, 00:10:50.004 { 00:10:50.004 "name": "BaseBdev3", 00:10:50.004 "uuid": "0d02744e-2130-42a1-b37f-0cab8b3471dc", 00:10:50.004 "is_configured": true, 00:10:50.004 "data_offset": 0, 00:10:50.004 "data_size": 65536 00:10:50.004 }, 00:10:50.004 { 00:10:50.004 "name": "BaseBdev4", 00:10:50.004 "uuid": "a5690c56-f60f-4d9e-8387-0fdcf8017f70", 00:10:50.004 "is_configured": true, 00:10:50.004 "data_offset": 0, 00:10:50.004 "data_size": 65536 00:10:50.004 } 00:10:50.004 ] 00:10:50.004 }' 00:10:50.004 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:50.004 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.573 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:10:50.573 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:50.573 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.573 [2024-11-19 15:16:40.645026] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:50.573 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:50.573 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:50.573 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:50.573 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:50.573 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:50.573 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:50.573 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:50.573 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:50.573 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:50.573 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:50.573 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:50.573 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:50.573 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:50.573 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:50.573 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.573 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:50.573 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:50.573 "name": "Existed_Raid", 00:10:50.573 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:50.573 "strip_size_kb": 64, 00:10:50.573 "state": "configuring", 00:10:50.573 "raid_level": "raid0", 00:10:50.573 "superblock": false, 00:10:50.573 "num_base_bdevs": 4, 00:10:50.573 "num_base_bdevs_discovered": 2, 00:10:50.573 "num_base_bdevs_operational": 4, 00:10:50.573 "base_bdevs_list": [ 00:10:50.573 { 00:10:50.573 "name": "BaseBdev1", 00:10:50.573 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:50.573 "is_configured": false, 00:10:50.573 "data_offset": 0, 00:10:50.573 "data_size": 0 00:10:50.573 }, 00:10:50.573 { 00:10:50.573 "name": null, 00:10:50.573 "uuid": "11d501d0-1e66-41a3-95e7-cf12d4eb94ab", 00:10:50.573 "is_configured": false, 00:10:50.573 "data_offset": 0, 00:10:50.573 "data_size": 65536 00:10:50.573 }, 00:10:50.573 { 00:10:50.573 "name": "BaseBdev3", 00:10:50.573 "uuid": "0d02744e-2130-42a1-b37f-0cab8b3471dc", 00:10:50.573 "is_configured": true, 00:10:50.573 "data_offset": 0, 00:10:50.573 "data_size": 65536 00:10:50.573 }, 00:10:50.573 { 00:10:50.573 "name": "BaseBdev4", 00:10:50.573 "uuid": "a5690c56-f60f-4d9e-8387-0fdcf8017f70", 00:10:50.573 "is_configured": true, 00:10:50.573 "data_offset": 0, 00:10:50.573 "data_size": 65536 00:10:50.573 } 00:10:50.573 ] 00:10:50.573 }' 00:10:50.573 15:16:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:50.573 15:16:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.838 15:16:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:50.838 15:16:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:50.838 15:16:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:50.838 15:16:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.838 15:16:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:50.838 15:16:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:10:50.838 15:16:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:50.838 15:16:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:50.838 15:16:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.114 [2024-11-19 15:16:41.176926] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:51.114 BaseBdev1 00:10:51.114 15:16:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:51.114 15:16:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:10:51.114 15:16:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:10:51.114 15:16:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:51.114 15:16:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:51.114 15:16:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:51.114 15:16:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:51.114 15:16:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:51.114 15:16:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:51.114 15:16:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.114 15:16:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:51.114 15:16:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:51.114 15:16:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:51.114 15:16:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.114 [ 00:10:51.114 { 00:10:51.114 "name": "BaseBdev1", 00:10:51.114 "aliases": [ 00:10:51.114 "6cc3bfc4-5637-4728-abbc-43d72f3687f9" 00:10:51.114 ], 00:10:51.114 "product_name": "Malloc disk", 00:10:51.114 "block_size": 512, 00:10:51.114 "num_blocks": 65536, 00:10:51.114 "uuid": "6cc3bfc4-5637-4728-abbc-43d72f3687f9", 00:10:51.114 "assigned_rate_limits": { 00:10:51.114 "rw_ios_per_sec": 0, 00:10:51.114 "rw_mbytes_per_sec": 0, 00:10:51.114 "r_mbytes_per_sec": 0, 00:10:51.114 "w_mbytes_per_sec": 0 00:10:51.114 }, 00:10:51.114 "claimed": true, 00:10:51.114 "claim_type": "exclusive_write", 00:10:51.114 "zoned": false, 00:10:51.114 "supported_io_types": { 00:10:51.114 "read": true, 00:10:51.114 "write": true, 00:10:51.114 "unmap": true, 00:10:51.114 "flush": true, 00:10:51.114 "reset": true, 00:10:51.114 "nvme_admin": false, 00:10:51.114 "nvme_io": false, 00:10:51.114 "nvme_io_md": false, 00:10:51.114 "write_zeroes": true, 00:10:51.114 "zcopy": true, 00:10:51.114 "get_zone_info": false, 00:10:51.114 "zone_management": false, 00:10:51.114 "zone_append": false, 00:10:51.114 "compare": false, 00:10:51.114 "compare_and_write": false, 00:10:51.114 "abort": true, 00:10:51.114 "seek_hole": false, 00:10:51.114 "seek_data": false, 00:10:51.114 "copy": true, 00:10:51.114 "nvme_iov_md": false 00:10:51.114 }, 00:10:51.114 "memory_domains": [ 00:10:51.114 { 00:10:51.114 "dma_device_id": "system", 00:10:51.114 "dma_device_type": 1 00:10:51.114 }, 00:10:51.114 { 00:10:51.114 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:51.114 "dma_device_type": 2 00:10:51.114 } 00:10:51.114 ], 00:10:51.114 "driver_specific": {} 00:10:51.114 } 00:10:51.114 ] 00:10:51.114 15:16:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:51.114 15:16:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:51.114 15:16:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:51.114 15:16:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:51.114 15:16:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:51.114 15:16:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:51.114 15:16:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:51.114 15:16:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:51.114 15:16:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:51.114 15:16:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:51.114 15:16:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:51.114 15:16:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:51.114 15:16:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:51.114 15:16:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:51.114 15:16:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.114 15:16:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:51.114 15:16:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:51.114 15:16:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:51.114 "name": "Existed_Raid", 00:10:51.114 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:51.114 "strip_size_kb": 64, 00:10:51.114 "state": "configuring", 00:10:51.114 "raid_level": "raid0", 00:10:51.114 "superblock": false, 00:10:51.114 "num_base_bdevs": 4, 00:10:51.114 "num_base_bdevs_discovered": 3, 00:10:51.114 "num_base_bdevs_operational": 4, 00:10:51.115 "base_bdevs_list": [ 00:10:51.115 { 00:10:51.115 "name": "BaseBdev1", 00:10:51.115 "uuid": "6cc3bfc4-5637-4728-abbc-43d72f3687f9", 00:10:51.115 "is_configured": true, 00:10:51.115 "data_offset": 0, 00:10:51.115 "data_size": 65536 00:10:51.115 }, 00:10:51.115 { 00:10:51.115 "name": null, 00:10:51.115 "uuid": "11d501d0-1e66-41a3-95e7-cf12d4eb94ab", 00:10:51.115 "is_configured": false, 00:10:51.115 "data_offset": 0, 00:10:51.115 "data_size": 65536 00:10:51.115 }, 00:10:51.115 { 00:10:51.115 "name": "BaseBdev3", 00:10:51.115 "uuid": "0d02744e-2130-42a1-b37f-0cab8b3471dc", 00:10:51.115 "is_configured": true, 00:10:51.115 "data_offset": 0, 00:10:51.115 "data_size": 65536 00:10:51.115 }, 00:10:51.115 { 00:10:51.115 "name": "BaseBdev4", 00:10:51.115 "uuid": "a5690c56-f60f-4d9e-8387-0fdcf8017f70", 00:10:51.115 "is_configured": true, 00:10:51.115 "data_offset": 0, 00:10:51.115 "data_size": 65536 00:10:51.115 } 00:10:51.115 ] 00:10:51.115 }' 00:10:51.115 15:16:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:51.115 15:16:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.374 15:16:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:51.374 15:16:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:51.374 15:16:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:51.374 15:16:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.374 15:16:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:51.374 15:16:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:10:51.374 15:16:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:10:51.374 15:16:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:51.374 15:16:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.634 [2024-11-19 15:16:41.716093] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:51.634 15:16:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:51.634 15:16:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:51.634 15:16:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:51.634 15:16:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:51.634 15:16:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:51.634 15:16:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:51.634 15:16:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:51.634 15:16:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:51.634 15:16:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:51.634 15:16:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:51.634 15:16:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:51.634 15:16:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:51.634 15:16:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:51.634 15:16:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:51.634 15:16:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.634 15:16:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:51.634 15:16:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:51.634 "name": "Existed_Raid", 00:10:51.634 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:51.634 "strip_size_kb": 64, 00:10:51.634 "state": "configuring", 00:10:51.634 "raid_level": "raid0", 00:10:51.634 "superblock": false, 00:10:51.634 "num_base_bdevs": 4, 00:10:51.634 "num_base_bdevs_discovered": 2, 00:10:51.634 "num_base_bdevs_operational": 4, 00:10:51.635 "base_bdevs_list": [ 00:10:51.635 { 00:10:51.635 "name": "BaseBdev1", 00:10:51.635 "uuid": "6cc3bfc4-5637-4728-abbc-43d72f3687f9", 00:10:51.635 "is_configured": true, 00:10:51.635 "data_offset": 0, 00:10:51.635 "data_size": 65536 00:10:51.635 }, 00:10:51.635 { 00:10:51.635 "name": null, 00:10:51.635 "uuid": "11d501d0-1e66-41a3-95e7-cf12d4eb94ab", 00:10:51.635 "is_configured": false, 00:10:51.635 "data_offset": 0, 00:10:51.635 "data_size": 65536 00:10:51.635 }, 00:10:51.635 { 00:10:51.635 "name": null, 00:10:51.635 "uuid": "0d02744e-2130-42a1-b37f-0cab8b3471dc", 00:10:51.635 "is_configured": false, 00:10:51.635 "data_offset": 0, 00:10:51.635 "data_size": 65536 00:10:51.635 }, 00:10:51.635 { 00:10:51.635 "name": "BaseBdev4", 00:10:51.635 "uuid": "a5690c56-f60f-4d9e-8387-0fdcf8017f70", 00:10:51.635 "is_configured": true, 00:10:51.635 "data_offset": 0, 00:10:51.635 "data_size": 65536 00:10:51.635 } 00:10:51.635 ] 00:10:51.635 }' 00:10:51.635 15:16:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:51.635 15:16:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.894 15:16:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:51.894 15:16:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:51.894 15:16:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:51.894 15:16:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.894 15:16:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:51.894 15:16:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:10:51.894 15:16:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:10:51.894 15:16:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:51.894 15:16:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.894 [2024-11-19 15:16:42.219461] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:51.894 15:16:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:51.894 15:16:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:51.894 15:16:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:51.894 15:16:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:51.894 15:16:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:51.894 15:16:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:51.894 15:16:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:51.894 15:16:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:51.894 15:16:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:51.894 15:16:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:51.894 15:16:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:51.894 15:16:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:51.894 15:16:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:52.153 15:16:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:52.153 15:16:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:52.153 15:16:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:52.153 15:16:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:52.153 "name": "Existed_Raid", 00:10:52.153 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:52.153 "strip_size_kb": 64, 00:10:52.153 "state": "configuring", 00:10:52.153 "raid_level": "raid0", 00:10:52.153 "superblock": false, 00:10:52.153 "num_base_bdevs": 4, 00:10:52.153 "num_base_bdevs_discovered": 3, 00:10:52.153 "num_base_bdevs_operational": 4, 00:10:52.153 "base_bdevs_list": [ 00:10:52.153 { 00:10:52.153 "name": "BaseBdev1", 00:10:52.153 "uuid": "6cc3bfc4-5637-4728-abbc-43d72f3687f9", 00:10:52.153 "is_configured": true, 00:10:52.153 "data_offset": 0, 00:10:52.153 "data_size": 65536 00:10:52.153 }, 00:10:52.153 { 00:10:52.153 "name": null, 00:10:52.153 "uuid": "11d501d0-1e66-41a3-95e7-cf12d4eb94ab", 00:10:52.153 "is_configured": false, 00:10:52.153 "data_offset": 0, 00:10:52.153 "data_size": 65536 00:10:52.153 }, 00:10:52.153 { 00:10:52.153 "name": "BaseBdev3", 00:10:52.153 "uuid": "0d02744e-2130-42a1-b37f-0cab8b3471dc", 00:10:52.153 "is_configured": true, 00:10:52.153 "data_offset": 0, 00:10:52.153 "data_size": 65536 00:10:52.153 }, 00:10:52.153 { 00:10:52.153 "name": "BaseBdev4", 00:10:52.153 "uuid": "a5690c56-f60f-4d9e-8387-0fdcf8017f70", 00:10:52.153 "is_configured": true, 00:10:52.153 "data_offset": 0, 00:10:52.153 "data_size": 65536 00:10:52.153 } 00:10:52.153 ] 00:10:52.153 }' 00:10:52.153 15:16:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:52.153 15:16:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:52.413 15:16:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:52.413 15:16:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:52.413 15:16:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:52.413 15:16:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:52.413 15:16:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:52.413 15:16:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:10:52.413 15:16:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:52.413 15:16:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:52.413 15:16:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:52.413 [2024-11-19 15:16:42.678798] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:52.413 15:16:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:52.413 15:16:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:52.413 15:16:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:52.413 15:16:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:52.413 15:16:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:52.413 15:16:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:52.413 15:16:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:52.413 15:16:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:52.413 15:16:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:52.414 15:16:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:52.414 15:16:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:52.414 15:16:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:52.414 15:16:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:52.414 15:16:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:52.414 15:16:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:52.414 15:16:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:52.414 15:16:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:52.414 "name": "Existed_Raid", 00:10:52.414 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:52.414 "strip_size_kb": 64, 00:10:52.414 "state": "configuring", 00:10:52.414 "raid_level": "raid0", 00:10:52.414 "superblock": false, 00:10:52.414 "num_base_bdevs": 4, 00:10:52.414 "num_base_bdevs_discovered": 2, 00:10:52.414 "num_base_bdevs_operational": 4, 00:10:52.414 "base_bdevs_list": [ 00:10:52.414 { 00:10:52.414 "name": null, 00:10:52.414 "uuid": "6cc3bfc4-5637-4728-abbc-43d72f3687f9", 00:10:52.414 "is_configured": false, 00:10:52.414 "data_offset": 0, 00:10:52.414 "data_size": 65536 00:10:52.414 }, 00:10:52.414 { 00:10:52.414 "name": null, 00:10:52.414 "uuid": "11d501d0-1e66-41a3-95e7-cf12d4eb94ab", 00:10:52.414 "is_configured": false, 00:10:52.414 "data_offset": 0, 00:10:52.414 "data_size": 65536 00:10:52.414 }, 00:10:52.414 { 00:10:52.414 "name": "BaseBdev3", 00:10:52.414 "uuid": "0d02744e-2130-42a1-b37f-0cab8b3471dc", 00:10:52.414 "is_configured": true, 00:10:52.414 "data_offset": 0, 00:10:52.414 "data_size": 65536 00:10:52.414 }, 00:10:52.414 { 00:10:52.414 "name": "BaseBdev4", 00:10:52.414 "uuid": "a5690c56-f60f-4d9e-8387-0fdcf8017f70", 00:10:52.414 "is_configured": true, 00:10:52.414 "data_offset": 0, 00:10:52.414 "data_size": 65536 00:10:52.414 } 00:10:52.414 ] 00:10:52.414 }' 00:10:52.414 15:16:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:52.414 15:16:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:52.982 15:16:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:52.982 15:16:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:52.982 15:16:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:52.982 15:16:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:52.982 15:16:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:52.982 15:16:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:10:52.982 15:16:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:10:52.982 15:16:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:52.982 15:16:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:52.982 [2024-11-19 15:16:43.201647] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:52.982 15:16:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:52.982 15:16:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:52.982 15:16:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:52.982 15:16:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:52.982 15:16:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:52.982 15:16:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:52.982 15:16:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:52.982 15:16:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:52.982 15:16:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:52.982 15:16:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:52.982 15:16:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:52.982 15:16:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:52.982 15:16:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:52.982 15:16:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:52.982 15:16:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:52.982 15:16:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:52.982 15:16:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:52.982 "name": "Existed_Raid", 00:10:52.982 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:52.982 "strip_size_kb": 64, 00:10:52.982 "state": "configuring", 00:10:52.982 "raid_level": "raid0", 00:10:52.982 "superblock": false, 00:10:52.982 "num_base_bdevs": 4, 00:10:52.982 "num_base_bdevs_discovered": 3, 00:10:52.982 "num_base_bdevs_operational": 4, 00:10:52.982 "base_bdevs_list": [ 00:10:52.982 { 00:10:52.982 "name": null, 00:10:52.982 "uuid": "6cc3bfc4-5637-4728-abbc-43d72f3687f9", 00:10:52.982 "is_configured": false, 00:10:52.982 "data_offset": 0, 00:10:52.982 "data_size": 65536 00:10:52.982 }, 00:10:52.982 { 00:10:52.982 "name": "BaseBdev2", 00:10:52.982 "uuid": "11d501d0-1e66-41a3-95e7-cf12d4eb94ab", 00:10:52.982 "is_configured": true, 00:10:52.982 "data_offset": 0, 00:10:52.982 "data_size": 65536 00:10:52.982 }, 00:10:52.982 { 00:10:52.982 "name": "BaseBdev3", 00:10:52.982 "uuid": "0d02744e-2130-42a1-b37f-0cab8b3471dc", 00:10:52.982 "is_configured": true, 00:10:52.982 "data_offset": 0, 00:10:52.982 "data_size": 65536 00:10:52.982 }, 00:10:52.982 { 00:10:52.982 "name": "BaseBdev4", 00:10:52.982 "uuid": "a5690c56-f60f-4d9e-8387-0fdcf8017f70", 00:10:52.982 "is_configured": true, 00:10:52.982 "data_offset": 0, 00:10:52.982 "data_size": 65536 00:10:52.982 } 00:10:52.982 ] 00:10:52.982 }' 00:10:52.982 15:16:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:52.982 15:16:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.551 15:16:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:53.551 15:16:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:53.551 15:16:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.551 15:16:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:53.551 15:16:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:53.551 15:16:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:10:53.551 15:16:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:53.551 15:16:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:53.551 15:16:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.551 15:16:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:10:53.551 15:16:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:53.551 15:16:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 6cc3bfc4-5637-4728-abbc-43d72f3687f9 00:10:53.551 15:16:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:53.551 15:16:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.551 [2024-11-19 15:16:43.781384] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:10:53.551 [2024-11-19 15:16:43.781436] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:10:53.551 [2024-11-19 15:16:43.781443] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:10:53.551 [2024-11-19 15:16:43.781741] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:10:53.551 [2024-11-19 15:16:43.781871] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:10:53.551 [2024-11-19 15:16:43.781883] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:10:53.551 [2024-11-19 15:16:43.782095] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:53.551 NewBaseBdev 00:10:53.551 15:16:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:53.551 15:16:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:10:53.551 15:16:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:10:53.551 15:16:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:53.551 15:16:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:53.551 15:16:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:53.551 15:16:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:53.551 15:16:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:53.551 15:16:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:53.551 15:16:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.551 15:16:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:53.552 15:16:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:10:53.552 15:16:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:53.552 15:16:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.552 [ 00:10:53.552 { 00:10:53.552 "name": "NewBaseBdev", 00:10:53.552 "aliases": [ 00:10:53.552 "6cc3bfc4-5637-4728-abbc-43d72f3687f9" 00:10:53.552 ], 00:10:53.552 "product_name": "Malloc disk", 00:10:53.552 "block_size": 512, 00:10:53.552 "num_blocks": 65536, 00:10:53.552 "uuid": "6cc3bfc4-5637-4728-abbc-43d72f3687f9", 00:10:53.552 "assigned_rate_limits": { 00:10:53.552 "rw_ios_per_sec": 0, 00:10:53.552 "rw_mbytes_per_sec": 0, 00:10:53.552 "r_mbytes_per_sec": 0, 00:10:53.552 "w_mbytes_per_sec": 0 00:10:53.552 }, 00:10:53.552 "claimed": true, 00:10:53.552 "claim_type": "exclusive_write", 00:10:53.552 "zoned": false, 00:10:53.552 "supported_io_types": { 00:10:53.552 "read": true, 00:10:53.552 "write": true, 00:10:53.552 "unmap": true, 00:10:53.552 "flush": true, 00:10:53.552 "reset": true, 00:10:53.552 "nvme_admin": false, 00:10:53.552 "nvme_io": false, 00:10:53.552 "nvme_io_md": false, 00:10:53.552 "write_zeroes": true, 00:10:53.552 "zcopy": true, 00:10:53.552 "get_zone_info": false, 00:10:53.552 "zone_management": false, 00:10:53.552 "zone_append": false, 00:10:53.552 "compare": false, 00:10:53.552 "compare_and_write": false, 00:10:53.552 "abort": true, 00:10:53.552 "seek_hole": false, 00:10:53.552 "seek_data": false, 00:10:53.552 "copy": true, 00:10:53.552 "nvme_iov_md": false 00:10:53.552 }, 00:10:53.552 "memory_domains": [ 00:10:53.552 { 00:10:53.552 "dma_device_id": "system", 00:10:53.552 "dma_device_type": 1 00:10:53.552 }, 00:10:53.552 { 00:10:53.552 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:53.552 "dma_device_type": 2 00:10:53.552 } 00:10:53.552 ], 00:10:53.552 "driver_specific": {} 00:10:53.552 } 00:10:53.552 ] 00:10:53.552 15:16:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:53.552 15:16:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:53.552 15:16:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:10:53.552 15:16:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:53.552 15:16:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:53.552 15:16:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:53.552 15:16:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:53.552 15:16:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:53.552 15:16:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:53.552 15:16:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:53.552 15:16:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:53.552 15:16:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:53.552 15:16:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:53.552 15:16:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:53.552 15:16:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.552 15:16:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:53.552 15:16:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:53.552 15:16:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:53.552 "name": "Existed_Raid", 00:10:53.552 "uuid": "7538398e-a52f-4bb7-8847-f47ff913ded2", 00:10:53.552 "strip_size_kb": 64, 00:10:53.552 "state": "online", 00:10:53.552 "raid_level": "raid0", 00:10:53.552 "superblock": false, 00:10:53.552 "num_base_bdevs": 4, 00:10:53.552 "num_base_bdevs_discovered": 4, 00:10:53.552 "num_base_bdevs_operational": 4, 00:10:53.552 "base_bdevs_list": [ 00:10:53.552 { 00:10:53.552 "name": "NewBaseBdev", 00:10:53.552 "uuid": "6cc3bfc4-5637-4728-abbc-43d72f3687f9", 00:10:53.552 "is_configured": true, 00:10:53.552 "data_offset": 0, 00:10:53.552 "data_size": 65536 00:10:53.552 }, 00:10:53.552 { 00:10:53.552 "name": "BaseBdev2", 00:10:53.552 "uuid": "11d501d0-1e66-41a3-95e7-cf12d4eb94ab", 00:10:53.552 "is_configured": true, 00:10:53.552 "data_offset": 0, 00:10:53.552 "data_size": 65536 00:10:53.552 }, 00:10:53.552 { 00:10:53.552 "name": "BaseBdev3", 00:10:53.552 "uuid": "0d02744e-2130-42a1-b37f-0cab8b3471dc", 00:10:53.552 "is_configured": true, 00:10:53.552 "data_offset": 0, 00:10:53.552 "data_size": 65536 00:10:53.552 }, 00:10:53.552 { 00:10:53.552 "name": "BaseBdev4", 00:10:53.552 "uuid": "a5690c56-f60f-4d9e-8387-0fdcf8017f70", 00:10:53.552 "is_configured": true, 00:10:53.552 "data_offset": 0, 00:10:53.552 "data_size": 65536 00:10:53.552 } 00:10:53.552 ] 00:10:53.552 }' 00:10:53.552 15:16:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:53.552 15:16:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.120 15:16:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:10:54.120 15:16:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:54.120 15:16:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:54.120 15:16:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:54.120 15:16:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:54.120 15:16:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:54.121 15:16:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:54.121 15:16:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.121 15:16:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.121 15:16:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:54.121 [2024-11-19 15:16:44.296926] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:54.121 15:16:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:54.121 15:16:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:54.121 "name": "Existed_Raid", 00:10:54.121 "aliases": [ 00:10:54.121 "7538398e-a52f-4bb7-8847-f47ff913ded2" 00:10:54.121 ], 00:10:54.121 "product_name": "Raid Volume", 00:10:54.121 "block_size": 512, 00:10:54.121 "num_blocks": 262144, 00:10:54.121 "uuid": "7538398e-a52f-4bb7-8847-f47ff913ded2", 00:10:54.121 "assigned_rate_limits": { 00:10:54.121 "rw_ios_per_sec": 0, 00:10:54.121 "rw_mbytes_per_sec": 0, 00:10:54.121 "r_mbytes_per_sec": 0, 00:10:54.121 "w_mbytes_per_sec": 0 00:10:54.121 }, 00:10:54.121 "claimed": false, 00:10:54.121 "zoned": false, 00:10:54.121 "supported_io_types": { 00:10:54.121 "read": true, 00:10:54.121 "write": true, 00:10:54.121 "unmap": true, 00:10:54.121 "flush": true, 00:10:54.121 "reset": true, 00:10:54.121 "nvme_admin": false, 00:10:54.121 "nvme_io": false, 00:10:54.121 "nvme_io_md": false, 00:10:54.121 "write_zeroes": true, 00:10:54.121 "zcopy": false, 00:10:54.121 "get_zone_info": false, 00:10:54.121 "zone_management": false, 00:10:54.121 "zone_append": false, 00:10:54.121 "compare": false, 00:10:54.121 "compare_and_write": false, 00:10:54.121 "abort": false, 00:10:54.121 "seek_hole": false, 00:10:54.121 "seek_data": false, 00:10:54.121 "copy": false, 00:10:54.121 "nvme_iov_md": false 00:10:54.121 }, 00:10:54.121 "memory_domains": [ 00:10:54.121 { 00:10:54.121 "dma_device_id": "system", 00:10:54.121 "dma_device_type": 1 00:10:54.121 }, 00:10:54.121 { 00:10:54.121 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:54.121 "dma_device_type": 2 00:10:54.121 }, 00:10:54.121 { 00:10:54.121 "dma_device_id": "system", 00:10:54.121 "dma_device_type": 1 00:10:54.121 }, 00:10:54.121 { 00:10:54.121 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:54.121 "dma_device_type": 2 00:10:54.121 }, 00:10:54.121 { 00:10:54.121 "dma_device_id": "system", 00:10:54.121 "dma_device_type": 1 00:10:54.121 }, 00:10:54.121 { 00:10:54.121 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:54.121 "dma_device_type": 2 00:10:54.121 }, 00:10:54.121 { 00:10:54.121 "dma_device_id": "system", 00:10:54.121 "dma_device_type": 1 00:10:54.121 }, 00:10:54.121 { 00:10:54.121 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:54.121 "dma_device_type": 2 00:10:54.121 } 00:10:54.121 ], 00:10:54.121 "driver_specific": { 00:10:54.121 "raid": { 00:10:54.121 "uuid": "7538398e-a52f-4bb7-8847-f47ff913ded2", 00:10:54.121 "strip_size_kb": 64, 00:10:54.121 "state": "online", 00:10:54.121 "raid_level": "raid0", 00:10:54.121 "superblock": false, 00:10:54.121 "num_base_bdevs": 4, 00:10:54.121 "num_base_bdevs_discovered": 4, 00:10:54.121 "num_base_bdevs_operational": 4, 00:10:54.121 "base_bdevs_list": [ 00:10:54.121 { 00:10:54.121 "name": "NewBaseBdev", 00:10:54.121 "uuid": "6cc3bfc4-5637-4728-abbc-43d72f3687f9", 00:10:54.121 "is_configured": true, 00:10:54.121 "data_offset": 0, 00:10:54.121 "data_size": 65536 00:10:54.121 }, 00:10:54.121 { 00:10:54.121 "name": "BaseBdev2", 00:10:54.121 "uuid": "11d501d0-1e66-41a3-95e7-cf12d4eb94ab", 00:10:54.121 "is_configured": true, 00:10:54.121 "data_offset": 0, 00:10:54.121 "data_size": 65536 00:10:54.121 }, 00:10:54.121 { 00:10:54.121 "name": "BaseBdev3", 00:10:54.121 "uuid": "0d02744e-2130-42a1-b37f-0cab8b3471dc", 00:10:54.121 "is_configured": true, 00:10:54.121 "data_offset": 0, 00:10:54.121 "data_size": 65536 00:10:54.121 }, 00:10:54.121 { 00:10:54.121 "name": "BaseBdev4", 00:10:54.121 "uuid": "a5690c56-f60f-4d9e-8387-0fdcf8017f70", 00:10:54.121 "is_configured": true, 00:10:54.121 "data_offset": 0, 00:10:54.121 "data_size": 65536 00:10:54.121 } 00:10:54.121 ] 00:10:54.121 } 00:10:54.121 } 00:10:54.121 }' 00:10:54.121 15:16:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:54.121 15:16:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:10:54.121 BaseBdev2 00:10:54.121 BaseBdev3 00:10:54.121 BaseBdev4' 00:10:54.121 15:16:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:54.121 15:16:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:54.121 15:16:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:54.121 15:16:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:10:54.121 15:16:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.121 15:16:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.121 15:16:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:54.121 15:16:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:54.381 15:16:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:54.381 15:16:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:54.381 15:16:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:54.381 15:16:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:54.381 15:16:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:54.381 15:16:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.381 15:16:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.381 15:16:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:54.381 15:16:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:54.381 15:16:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:54.381 15:16:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:54.381 15:16:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:54.381 15:16:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:54.381 15:16:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.381 15:16:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.381 15:16:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:54.381 15:16:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:54.381 15:16:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:54.381 15:16:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:54.381 15:16:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:54.381 15:16:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:54.381 15:16:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.381 15:16:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.381 15:16:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:54.381 15:16:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:54.381 15:16:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:54.381 15:16:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:54.381 15:16:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.381 15:16:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.381 [2024-11-19 15:16:44.616034] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:54.381 [2024-11-19 15:16:44.616079] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:54.381 [2024-11-19 15:16:44.616166] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:54.381 [2024-11-19 15:16:44.616244] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:54.381 [2024-11-19 15:16:44.616268] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:10:54.381 15:16:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:54.381 15:16:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 80356 00:10:54.381 15:16:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 80356 ']' 00:10:54.381 15:16:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 80356 00:10:54.381 15:16:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:10:54.381 15:16:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:54.381 15:16:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 80356 00:10:54.381 killing process with pid 80356 00:10:54.381 15:16:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:54.381 15:16:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:54.381 15:16:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 80356' 00:10:54.381 15:16:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 80356 00:10:54.381 [2024-11-19 15:16:44.648996] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:54.382 15:16:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 80356 00:10:54.641 [2024-11-19 15:16:44.726828] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:54.901 15:16:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:10:54.901 00:10:54.901 real 0m9.885s 00:10:54.901 user 0m16.698s 00:10:54.901 sys 0m2.077s 00:10:54.901 15:16:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:54.901 15:16:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.901 ************************************ 00:10:54.901 END TEST raid_state_function_test 00:10:54.901 ************************************ 00:10:54.901 15:16:45 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid0 4 true 00:10:54.901 15:16:45 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:10:54.901 15:16:45 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:54.901 15:16:45 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:54.901 ************************************ 00:10:54.901 START TEST raid_state_function_test_sb 00:10:54.901 ************************************ 00:10:54.901 15:16:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid0 4 true 00:10:54.901 15:16:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:10:54.901 15:16:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:10:54.901 15:16:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:10:54.901 15:16:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:10:54.901 15:16:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:10:54.901 15:16:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:54.901 15:16:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:10:54.901 15:16:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:54.901 15:16:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:54.901 15:16:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:10:54.901 15:16:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:54.901 15:16:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:54.901 15:16:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:10:54.901 15:16:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:54.901 15:16:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:54.901 15:16:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:10:54.901 15:16:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:54.901 15:16:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:54.901 15:16:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:54.901 15:16:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:10:54.901 15:16:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:10:54.901 15:16:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:10:54.901 15:16:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:10:54.901 15:16:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:10:54.901 15:16:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:10:54.901 15:16:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:10:54.901 15:16:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:10:54.901 15:16:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:10:54.901 15:16:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:10:54.901 15:16:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=81016 00:10:54.901 15:16:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:10:54.901 15:16:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 81016' 00:10:54.901 Process raid pid: 81016 00:10:54.902 15:16:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 81016 00:10:54.902 15:16:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 81016 ']' 00:10:54.902 15:16:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:54.902 15:16:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:54.902 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:54.902 15:16:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:54.902 15:16:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:54.902 15:16:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:54.902 [2024-11-19 15:16:45.201215] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:10:54.902 [2024-11-19 15:16:45.201494] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:55.164 [2024-11-19 15:16:45.354994] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:55.164 [2024-11-19 15:16:45.394284] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:55.164 [2024-11-19 15:16:45.470169] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:55.164 [2024-11-19 15:16:45.470205] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:55.737 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:55.737 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:10:55.737 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:55.737 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.737 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:55.737 [2024-11-19 15:16:46.029569] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:55.737 [2024-11-19 15:16:46.029635] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:55.737 [2024-11-19 15:16:46.029645] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:55.737 [2024-11-19 15:16:46.029657] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:55.737 [2024-11-19 15:16:46.029663] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:55.737 [2024-11-19 15:16:46.029675] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:55.737 [2024-11-19 15:16:46.029680] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:55.737 [2024-11-19 15:16:46.029690] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:55.737 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.737 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:55.737 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:55.737 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:55.737 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:55.737 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:55.738 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:55.738 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:55.738 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:55.738 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:55.738 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:55.738 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:55.738 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.738 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:55.738 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:55.738 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.998 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:55.998 "name": "Existed_Raid", 00:10:55.998 "uuid": "04008d93-61d4-4849-bde5-70a35a7a7921", 00:10:55.998 "strip_size_kb": 64, 00:10:55.998 "state": "configuring", 00:10:55.998 "raid_level": "raid0", 00:10:55.998 "superblock": true, 00:10:55.998 "num_base_bdevs": 4, 00:10:55.998 "num_base_bdevs_discovered": 0, 00:10:55.998 "num_base_bdevs_operational": 4, 00:10:55.998 "base_bdevs_list": [ 00:10:55.998 { 00:10:55.998 "name": "BaseBdev1", 00:10:55.998 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:55.998 "is_configured": false, 00:10:55.998 "data_offset": 0, 00:10:55.998 "data_size": 0 00:10:55.998 }, 00:10:55.998 { 00:10:55.998 "name": "BaseBdev2", 00:10:55.998 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:55.998 "is_configured": false, 00:10:55.998 "data_offset": 0, 00:10:55.998 "data_size": 0 00:10:55.998 }, 00:10:55.998 { 00:10:55.998 "name": "BaseBdev3", 00:10:55.998 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:55.998 "is_configured": false, 00:10:55.998 "data_offset": 0, 00:10:55.998 "data_size": 0 00:10:55.998 }, 00:10:55.998 { 00:10:55.998 "name": "BaseBdev4", 00:10:55.998 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:55.998 "is_configured": false, 00:10:55.998 "data_offset": 0, 00:10:55.998 "data_size": 0 00:10:55.998 } 00:10:55.998 ] 00:10:55.998 }' 00:10:55.998 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:55.998 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:56.258 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:56.258 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:56.258 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:56.258 [2024-11-19 15:16:46.436829] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:56.258 [2024-11-19 15:16:46.436888] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:10:56.258 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:56.258 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:56.258 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:56.258 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:56.258 [2024-11-19 15:16:46.448793] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:56.258 [2024-11-19 15:16:46.448839] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:56.258 [2024-11-19 15:16:46.448848] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:56.258 [2024-11-19 15:16:46.448858] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:56.258 [2024-11-19 15:16:46.448863] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:56.258 [2024-11-19 15:16:46.448873] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:56.258 [2024-11-19 15:16:46.448879] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:56.258 [2024-11-19 15:16:46.448889] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:56.258 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:56.258 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:56.258 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:56.258 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:56.258 [2024-11-19 15:16:46.475839] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:56.258 BaseBdev1 00:10:56.258 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:56.258 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:10:56.258 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:10:56.258 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:56.258 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:56.258 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:56.259 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:56.259 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:56.259 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:56.259 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:56.259 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:56.259 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:56.259 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:56.259 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:56.259 [ 00:10:56.259 { 00:10:56.259 "name": "BaseBdev1", 00:10:56.259 "aliases": [ 00:10:56.259 "fdc91caf-941a-4265-b629-eb67820313f9" 00:10:56.259 ], 00:10:56.259 "product_name": "Malloc disk", 00:10:56.259 "block_size": 512, 00:10:56.259 "num_blocks": 65536, 00:10:56.259 "uuid": "fdc91caf-941a-4265-b629-eb67820313f9", 00:10:56.259 "assigned_rate_limits": { 00:10:56.259 "rw_ios_per_sec": 0, 00:10:56.259 "rw_mbytes_per_sec": 0, 00:10:56.259 "r_mbytes_per_sec": 0, 00:10:56.259 "w_mbytes_per_sec": 0 00:10:56.259 }, 00:10:56.259 "claimed": true, 00:10:56.259 "claim_type": "exclusive_write", 00:10:56.259 "zoned": false, 00:10:56.259 "supported_io_types": { 00:10:56.259 "read": true, 00:10:56.259 "write": true, 00:10:56.259 "unmap": true, 00:10:56.259 "flush": true, 00:10:56.259 "reset": true, 00:10:56.259 "nvme_admin": false, 00:10:56.259 "nvme_io": false, 00:10:56.259 "nvme_io_md": false, 00:10:56.259 "write_zeroes": true, 00:10:56.259 "zcopy": true, 00:10:56.259 "get_zone_info": false, 00:10:56.259 "zone_management": false, 00:10:56.259 "zone_append": false, 00:10:56.259 "compare": false, 00:10:56.259 "compare_and_write": false, 00:10:56.259 "abort": true, 00:10:56.259 "seek_hole": false, 00:10:56.259 "seek_data": false, 00:10:56.259 "copy": true, 00:10:56.259 "nvme_iov_md": false 00:10:56.259 }, 00:10:56.259 "memory_domains": [ 00:10:56.259 { 00:10:56.259 "dma_device_id": "system", 00:10:56.259 "dma_device_type": 1 00:10:56.259 }, 00:10:56.259 { 00:10:56.259 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:56.259 "dma_device_type": 2 00:10:56.259 } 00:10:56.259 ], 00:10:56.259 "driver_specific": {} 00:10:56.259 } 00:10:56.259 ] 00:10:56.259 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:56.259 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:56.259 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:56.259 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:56.259 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:56.259 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:56.259 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:56.259 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:56.259 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:56.259 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:56.259 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:56.259 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:56.259 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:56.259 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:56.259 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:56.259 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:56.259 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:56.259 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:56.259 "name": "Existed_Raid", 00:10:56.259 "uuid": "aaead4d9-125e-48c4-9a1c-d59759380e64", 00:10:56.259 "strip_size_kb": 64, 00:10:56.259 "state": "configuring", 00:10:56.259 "raid_level": "raid0", 00:10:56.259 "superblock": true, 00:10:56.259 "num_base_bdevs": 4, 00:10:56.259 "num_base_bdevs_discovered": 1, 00:10:56.259 "num_base_bdevs_operational": 4, 00:10:56.259 "base_bdevs_list": [ 00:10:56.259 { 00:10:56.259 "name": "BaseBdev1", 00:10:56.259 "uuid": "fdc91caf-941a-4265-b629-eb67820313f9", 00:10:56.259 "is_configured": true, 00:10:56.259 "data_offset": 2048, 00:10:56.259 "data_size": 63488 00:10:56.259 }, 00:10:56.259 { 00:10:56.259 "name": "BaseBdev2", 00:10:56.259 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:56.259 "is_configured": false, 00:10:56.259 "data_offset": 0, 00:10:56.259 "data_size": 0 00:10:56.259 }, 00:10:56.259 { 00:10:56.259 "name": "BaseBdev3", 00:10:56.259 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:56.259 "is_configured": false, 00:10:56.259 "data_offset": 0, 00:10:56.259 "data_size": 0 00:10:56.259 }, 00:10:56.259 { 00:10:56.259 "name": "BaseBdev4", 00:10:56.259 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:56.259 "is_configured": false, 00:10:56.259 "data_offset": 0, 00:10:56.259 "data_size": 0 00:10:56.259 } 00:10:56.259 ] 00:10:56.259 }' 00:10:56.259 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:56.259 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:56.829 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:56.829 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:56.829 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:56.829 [2024-11-19 15:16:46.951141] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:56.829 [2024-11-19 15:16:46.951219] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:10:56.829 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:56.829 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:56.829 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:56.829 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:56.829 [2024-11-19 15:16:46.963143] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:56.829 [2024-11-19 15:16:46.965282] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:56.829 [2024-11-19 15:16:46.965321] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:56.829 [2024-11-19 15:16:46.965330] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:56.829 [2024-11-19 15:16:46.965339] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:56.829 [2024-11-19 15:16:46.965345] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:56.829 [2024-11-19 15:16:46.965353] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:56.829 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:56.829 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:10:56.829 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:56.829 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:56.829 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:56.829 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:56.829 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:56.829 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:56.829 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:56.829 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:56.829 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:56.829 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:56.829 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:56.829 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:56.829 15:16:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:56.829 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:56.829 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:56.829 15:16:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:56.829 15:16:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:56.829 "name": "Existed_Raid", 00:10:56.829 "uuid": "01f22662-bc99-4826-948f-2f5f386471f9", 00:10:56.829 "strip_size_kb": 64, 00:10:56.829 "state": "configuring", 00:10:56.829 "raid_level": "raid0", 00:10:56.829 "superblock": true, 00:10:56.829 "num_base_bdevs": 4, 00:10:56.829 "num_base_bdevs_discovered": 1, 00:10:56.829 "num_base_bdevs_operational": 4, 00:10:56.829 "base_bdevs_list": [ 00:10:56.829 { 00:10:56.829 "name": "BaseBdev1", 00:10:56.829 "uuid": "fdc91caf-941a-4265-b629-eb67820313f9", 00:10:56.829 "is_configured": true, 00:10:56.829 "data_offset": 2048, 00:10:56.829 "data_size": 63488 00:10:56.829 }, 00:10:56.829 { 00:10:56.829 "name": "BaseBdev2", 00:10:56.829 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:56.829 "is_configured": false, 00:10:56.829 "data_offset": 0, 00:10:56.829 "data_size": 0 00:10:56.829 }, 00:10:56.829 { 00:10:56.829 "name": "BaseBdev3", 00:10:56.829 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:56.829 "is_configured": false, 00:10:56.829 "data_offset": 0, 00:10:56.829 "data_size": 0 00:10:56.829 }, 00:10:56.829 { 00:10:56.829 "name": "BaseBdev4", 00:10:56.829 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:56.829 "is_configured": false, 00:10:56.829 "data_offset": 0, 00:10:56.829 "data_size": 0 00:10:56.829 } 00:10:56.829 ] 00:10:56.829 }' 00:10:56.829 15:16:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:56.829 15:16:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:57.090 15:16:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:57.090 15:16:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:57.090 15:16:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:57.090 [2024-11-19 15:16:47.415046] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:57.090 BaseBdev2 00:10:57.090 15:16:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:57.090 15:16:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:10:57.090 15:16:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:10:57.090 15:16:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:57.090 15:16:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:57.090 15:16:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:57.090 15:16:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:57.090 15:16:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:57.090 15:16:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:57.090 15:16:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:57.350 15:16:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:57.350 15:16:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:57.350 15:16:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:57.350 15:16:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:57.350 [ 00:10:57.350 { 00:10:57.350 "name": "BaseBdev2", 00:10:57.350 "aliases": [ 00:10:57.350 "34ba2329-9ba9-4605-8ef7-38847e6a5770" 00:10:57.350 ], 00:10:57.350 "product_name": "Malloc disk", 00:10:57.350 "block_size": 512, 00:10:57.350 "num_blocks": 65536, 00:10:57.350 "uuid": "34ba2329-9ba9-4605-8ef7-38847e6a5770", 00:10:57.351 "assigned_rate_limits": { 00:10:57.351 "rw_ios_per_sec": 0, 00:10:57.351 "rw_mbytes_per_sec": 0, 00:10:57.351 "r_mbytes_per_sec": 0, 00:10:57.351 "w_mbytes_per_sec": 0 00:10:57.351 }, 00:10:57.351 "claimed": true, 00:10:57.351 "claim_type": "exclusive_write", 00:10:57.351 "zoned": false, 00:10:57.351 "supported_io_types": { 00:10:57.351 "read": true, 00:10:57.351 "write": true, 00:10:57.351 "unmap": true, 00:10:57.351 "flush": true, 00:10:57.351 "reset": true, 00:10:57.351 "nvme_admin": false, 00:10:57.351 "nvme_io": false, 00:10:57.351 "nvme_io_md": false, 00:10:57.351 "write_zeroes": true, 00:10:57.351 "zcopy": true, 00:10:57.351 "get_zone_info": false, 00:10:57.351 "zone_management": false, 00:10:57.351 "zone_append": false, 00:10:57.351 "compare": false, 00:10:57.351 "compare_and_write": false, 00:10:57.351 "abort": true, 00:10:57.351 "seek_hole": false, 00:10:57.351 "seek_data": false, 00:10:57.351 "copy": true, 00:10:57.351 "nvme_iov_md": false 00:10:57.351 }, 00:10:57.351 "memory_domains": [ 00:10:57.351 { 00:10:57.351 "dma_device_id": "system", 00:10:57.351 "dma_device_type": 1 00:10:57.351 }, 00:10:57.351 { 00:10:57.351 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:57.351 "dma_device_type": 2 00:10:57.351 } 00:10:57.351 ], 00:10:57.351 "driver_specific": {} 00:10:57.351 } 00:10:57.351 ] 00:10:57.351 15:16:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:57.351 15:16:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:57.351 15:16:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:57.351 15:16:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:57.351 15:16:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:57.351 15:16:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:57.351 15:16:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:57.351 15:16:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:57.351 15:16:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:57.351 15:16:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:57.351 15:16:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:57.351 15:16:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:57.351 15:16:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:57.351 15:16:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:57.351 15:16:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:57.351 15:16:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:57.351 15:16:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:57.351 15:16:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:57.351 15:16:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:57.351 15:16:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:57.351 "name": "Existed_Raid", 00:10:57.351 "uuid": "01f22662-bc99-4826-948f-2f5f386471f9", 00:10:57.351 "strip_size_kb": 64, 00:10:57.351 "state": "configuring", 00:10:57.351 "raid_level": "raid0", 00:10:57.351 "superblock": true, 00:10:57.351 "num_base_bdevs": 4, 00:10:57.351 "num_base_bdevs_discovered": 2, 00:10:57.351 "num_base_bdevs_operational": 4, 00:10:57.351 "base_bdevs_list": [ 00:10:57.351 { 00:10:57.351 "name": "BaseBdev1", 00:10:57.351 "uuid": "fdc91caf-941a-4265-b629-eb67820313f9", 00:10:57.351 "is_configured": true, 00:10:57.351 "data_offset": 2048, 00:10:57.351 "data_size": 63488 00:10:57.351 }, 00:10:57.351 { 00:10:57.351 "name": "BaseBdev2", 00:10:57.351 "uuid": "34ba2329-9ba9-4605-8ef7-38847e6a5770", 00:10:57.351 "is_configured": true, 00:10:57.351 "data_offset": 2048, 00:10:57.351 "data_size": 63488 00:10:57.351 }, 00:10:57.351 { 00:10:57.351 "name": "BaseBdev3", 00:10:57.351 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:57.351 "is_configured": false, 00:10:57.351 "data_offset": 0, 00:10:57.351 "data_size": 0 00:10:57.351 }, 00:10:57.351 { 00:10:57.351 "name": "BaseBdev4", 00:10:57.351 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:57.351 "is_configured": false, 00:10:57.351 "data_offset": 0, 00:10:57.351 "data_size": 0 00:10:57.351 } 00:10:57.351 ] 00:10:57.351 }' 00:10:57.351 15:16:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:57.351 15:16:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:57.611 15:16:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:57.611 15:16:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:57.611 15:16:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:57.611 [2024-11-19 15:16:47.940173] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:57.611 BaseBdev3 00:10:57.611 15:16:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:57.611 15:16:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:10:57.611 15:16:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:10:57.611 15:16:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:57.611 15:16:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:57.611 15:16:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:57.611 15:16:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:57.611 15:16:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:57.611 15:16:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:57.611 15:16:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:57.872 15:16:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:57.872 15:16:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:57.872 15:16:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:57.872 15:16:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:57.872 [ 00:10:57.872 { 00:10:57.872 "name": "BaseBdev3", 00:10:57.872 "aliases": [ 00:10:57.872 "1f7ffec7-a57f-4071-adae-3e2894b82c8c" 00:10:57.872 ], 00:10:57.872 "product_name": "Malloc disk", 00:10:57.872 "block_size": 512, 00:10:57.872 "num_blocks": 65536, 00:10:57.872 "uuid": "1f7ffec7-a57f-4071-adae-3e2894b82c8c", 00:10:57.872 "assigned_rate_limits": { 00:10:57.872 "rw_ios_per_sec": 0, 00:10:57.872 "rw_mbytes_per_sec": 0, 00:10:57.872 "r_mbytes_per_sec": 0, 00:10:57.872 "w_mbytes_per_sec": 0 00:10:57.872 }, 00:10:57.872 "claimed": true, 00:10:57.872 "claim_type": "exclusive_write", 00:10:57.872 "zoned": false, 00:10:57.872 "supported_io_types": { 00:10:57.872 "read": true, 00:10:57.872 "write": true, 00:10:57.872 "unmap": true, 00:10:57.872 "flush": true, 00:10:57.872 "reset": true, 00:10:57.872 "nvme_admin": false, 00:10:57.872 "nvme_io": false, 00:10:57.872 "nvme_io_md": false, 00:10:57.872 "write_zeroes": true, 00:10:57.872 "zcopy": true, 00:10:57.872 "get_zone_info": false, 00:10:57.872 "zone_management": false, 00:10:57.872 "zone_append": false, 00:10:57.872 "compare": false, 00:10:57.872 "compare_and_write": false, 00:10:57.872 "abort": true, 00:10:57.872 "seek_hole": false, 00:10:57.872 "seek_data": false, 00:10:57.872 "copy": true, 00:10:57.872 "nvme_iov_md": false 00:10:57.872 }, 00:10:57.872 "memory_domains": [ 00:10:57.872 { 00:10:57.872 "dma_device_id": "system", 00:10:57.872 "dma_device_type": 1 00:10:57.872 }, 00:10:57.872 { 00:10:57.872 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:57.872 "dma_device_type": 2 00:10:57.872 } 00:10:57.872 ], 00:10:57.872 "driver_specific": {} 00:10:57.872 } 00:10:57.872 ] 00:10:57.872 15:16:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:57.872 15:16:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:57.872 15:16:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:57.872 15:16:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:57.872 15:16:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:57.872 15:16:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:57.872 15:16:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:57.872 15:16:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:57.872 15:16:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:57.872 15:16:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:57.872 15:16:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:57.872 15:16:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:57.872 15:16:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:57.872 15:16:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:57.872 15:16:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:57.872 15:16:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:57.872 15:16:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:57.872 15:16:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:57.872 15:16:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:57.872 15:16:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:57.872 "name": "Existed_Raid", 00:10:57.872 "uuid": "01f22662-bc99-4826-948f-2f5f386471f9", 00:10:57.872 "strip_size_kb": 64, 00:10:57.872 "state": "configuring", 00:10:57.872 "raid_level": "raid0", 00:10:57.872 "superblock": true, 00:10:57.872 "num_base_bdevs": 4, 00:10:57.872 "num_base_bdevs_discovered": 3, 00:10:57.872 "num_base_bdevs_operational": 4, 00:10:57.872 "base_bdevs_list": [ 00:10:57.872 { 00:10:57.872 "name": "BaseBdev1", 00:10:57.872 "uuid": "fdc91caf-941a-4265-b629-eb67820313f9", 00:10:57.872 "is_configured": true, 00:10:57.872 "data_offset": 2048, 00:10:57.872 "data_size": 63488 00:10:57.872 }, 00:10:57.872 { 00:10:57.872 "name": "BaseBdev2", 00:10:57.872 "uuid": "34ba2329-9ba9-4605-8ef7-38847e6a5770", 00:10:57.872 "is_configured": true, 00:10:57.872 "data_offset": 2048, 00:10:57.872 "data_size": 63488 00:10:57.872 }, 00:10:57.872 { 00:10:57.872 "name": "BaseBdev3", 00:10:57.872 "uuid": "1f7ffec7-a57f-4071-adae-3e2894b82c8c", 00:10:57.872 "is_configured": true, 00:10:57.872 "data_offset": 2048, 00:10:57.872 "data_size": 63488 00:10:57.872 }, 00:10:57.872 { 00:10:57.872 "name": "BaseBdev4", 00:10:57.872 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:57.872 "is_configured": false, 00:10:57.872 "data_offset": 0, 00:10:57.872 "data_size": 0 00:10:57.872 } 00:10:57.872 ] 00:10:57.872 }' 00:10:57.872 15:16:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:57.872 15:16:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:58.133 15:16:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:58.133 15:16:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.133 15:16:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:58.133 [2024-11-19 15:16:48.448030] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:58.133 [2024-11-19 15:16:48.448257] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:10:58.133 [2024-11-19 15:16:48.448281] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:58.133 BaseBdev4 00:10:58.133 [2024-11-19 15:16:48.448652] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:10:58.133 [2024-11-19 15:16:48.448798] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:10:58.133 [2024-11-19 15:16:48.448817] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:10:58.133 [2024-11-19 15:16:48.448935] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:58.133 15:16:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.133 15:16:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:10:58.133 15:16:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:10:58.133 15:16:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:58.133 15:16:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:58.133 15:16:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:58.133 15:16:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:58.133 15:16:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:58.133 15:16:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.133 15:16:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:58.133 15:16:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.133 15:16:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:58.133 15:16:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.133 15:16:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:58.393 [ 00:10:58.393 { 00:10:58.393 "name": "BaseBdev4", 00:10:58.393 "aliases": [ 00:10:58.393 "d373aed6-8b75-4d58-8347-b3d5a22f6fc6" 00:10:58.393 ], 00:10:58.393 "product_name": "Malloc disk", 00:10:58.393 "block_size": 512, 00:10:58.393 "num_blocks": 65536, 00:10:58.393 "uuid": "d373aed6-8b75-4d58-8347-b3d5a22f6fc6", 00:10:58.393 "assigned_rate_limits": { 00:10:58.393 "rw_ios_per_sec": 0, 00:10:58.393 "rw_mbytes_per_sec": 0, 00:10:58.393 "r_mbytes_per_sec": 0, 00:10:58.393 "w_mbytes_per_sec": 0 00:10:58.393 }, 00:10:58.393 "claimed": true, 00:10:58.393 "claim_type": "exclusive_write", 00:10:58.393 "zoned": false, 00:10:58.394 "supported_io_types": { 00:10:58.394 "read": true, 00:10:58.394 "write": true, 00:10:58.394 "unmap": true, 00:10:58.394 "flush": true, 00:10:58.394 "reset": true, 00:10:58.394 "nvme_admin": false, 00:10:58.394 "nvme_io": false, 00:10:58.394 "nvme_io_md": false, 00:10:58.394 "write_zeroes": true, 00:10:58.394 "zcopy": true, 00:10:58.394 "get_zone_info": false, 00:10:58.394 "zone_management": false, 00:10:58.394 "zone_append": false, 00:10:58.394 "compare": false, 00:10:58.394 "compare_and_write": false, 00:10:58.394 "abort": true, 00:10:58.394 "seek_hole": false, 00:10:58.394 "seek_data": false, 00:10:58.394 "copy": true, 00:10:58.394 "nvme_iov_md": false 00:10:58.394 }, 00:10:58.394 "memory_domains": [ 00:10:58.394 { 00:10:58.394 "dma_device_id": "system", 00:10:58.394 "dma_device_type": 1 00:10:58.394 }, 00:10:58.394 { 00:10:58.394 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:58.394 "dma_device_type": 2 00:10:58.394 } 00:10:58.394 ], 00:10:58.394 "driver_specific": {} 00:10:58.394 } 00:10:58.394 ] 00:10:58.394 15:16:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.394 15:16:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:58.394 15:16:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:58.394 15:16:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:58.394 15:16:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:10:58.394 15:16:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:58.394 15:16:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:58.394 15:16:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:58.394 15:16:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:58.394 15:16:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:58.394 15:16:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:58.394 15:16:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:58.394 15:16:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:58.394 15:16:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:58.394 15:16:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:58.394 15:16:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.394 15:16:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:58.394 15:16:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:58.394 15:16:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.394 15:16:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:58.394 "name": "Existed_Raid", 00:10:58.394 "uuid": "01f22662-bc99-4826-948f-2f5f386471f9", 00:10:58.394 "strip_size_kb": 64, 00:10:58.394 "state": "online", 00:10:58.394 "raid_level": "raid0", 00:10:58.394 "superblock": true, 00:10:58.394 "num_base_bdevs": 4, 00:10:58.394 "num_base_bdevs_discovered": 4, 00:10:58.394 "num_base_bdevs_operational": 4, 00:10:58.394 "base_bdevs_list": [ 00:10:58.394 { 00:10:58.394 "name": "BaseBdev1", 00:10:58.394 "uuid": "fdc91caf-941a-4265-b629-eb67820313f9", 00:10:58.394 "is_configured": true, 00:10:58.394 "data_offset": 2048, 00:10:58.394 "data_size": 63488 00:10:58.394 }, 00:10:58.394 { 00:10:58.394 "name": "BaseBdev2", 00:10:58.394 "uuid": "34ba2329-9ba9-4605-8ef7-38847e6a5770", 00:10:58.394 "is_configured": true, 00:10:58.394 "data_offset": 2048, 00:10:58.394 "data_size": 63488 00:10:58.394 }, 00:10:58.394 { 00:10:58.394 "name": "BaseBdev3", 00:10:58.394 "uuid": "1f7ffec7-a57f-4071-adae-3e2894b82c8c", 00:10:58.394 "is_configured": true, 00:10:58.394 "data_offset": 2048, 00:10:58.394 "data_size": 63488 00:10:58.394 }, 00:10:58.394 { 00:10:58.394 "name": "BaseBdev4", 00:10:58.394 "uuid": "d373aed6-8b75-4d58-8347-b3d5a22f6fc6", 00:10:58.394 "is_configured": true, 00:10:58.394 "data_offset": 2048, 00:10:58.394 "data_size": 63488 00:10:58.394 } 00:10:58.394 ] 00:10:58.394 }' 00:10:58.394 15:16:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:58.394 15:16:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:58.654 15:16:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:10:58.654 15:16:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:58.654 15:16:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:58.654 15:16:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:58.654 15:16:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:10:58.654 15:16:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:58.654 15:16:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:58.654 15:16:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.654 15:16:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:58.654 15:16:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:58.654 [2024-11-19 15:16:48.979463] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:58.914 15:16:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.914 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:58.914 "name": "Existed_Raid", 00:10:58.914 "aliases": [ 00:10:58.914 "01f22662-bc99-4826-948f-2f5f386471f9" 00:10:58.914 ], 00:10:58.914 "product_name": "Raid Volume", 00:10:58.914 "block_size": 512, 00:10:58.914 "num_blocks": 253952, 00:10:58.914 "uuid": "01f22662-bc99-4826-948f-2f5f386471f9", 00:10:58.914 "assigned_rate_limits": { 00:10:58.914 "rw_ios_per_sec": 0, 00:10:58.914 "rw_mbytes_per_sec": 0, 00:10:58.914 "r_mbytes_per_sec": 0, 00:10:58.914 "w_mbytes_per_sec": 0 00:10:58.914 }, 00:10:58.914 "claimed": false, 00:10:58.914 "zoned": false, 00:10:58.914 "supported_io_types": { 00:10:58.914 "read": true, 00:10:58.914 "write": true, 00:10:58.914 "unmap": true, 00:10:58.914 "flush": true, 00:10:58.914 "reset": true, 00:10:58.914 "nvme_admin": false, 00:10:58.914 "nvme_io": false, 00:10:58.914 "nvme_io_md": false, 00:10:58.914 "write_zeroes": true, 00:10:58.914 "zcopy": false, 00:10:58.914 "get_zone_info": false, 00:10:58.914 "zone_management": false, 00:10:58.914 "zone_append": false, 00:10:58.914 "compare": false, 00:10:58.914 "compare_and_write": false, 00:10:58.914 "abort": false, 00:10:58.914 "seek_hole": false, 00:10:58.914 "seek_data": false, 00:10:58.914 "copy": false, 00:10:58.914 "nvme_iov_md": false 00:10:58.914 }, 00:10:58.914 "memory_domains": [ 00:10:58.914 { 00:10:58.914 "dma_device_id": "system", 00:10:58.914 "dma_device_type": 1 00:10:58.914 }, 00:10:58.914 { 00:10:58.914 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:58.914 "dma_device_type": 2 00:10:58.914 }, 00:10:58.914 { 00:10:58.914 "dma_device_id": "system", 00:10:58.914 "dma_device_type": 1 00:10:58.914 }, 00:10:58.914 { 00:10:58.914 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:58.915 "dma_device_type": 2 00:10:58.915 }, 00:10:58.915 { 00:10:58.915 "dma_device_id": "system", 00:10:58.915 "dma_device_type": 1 00:10:58.915 }, 00:10:58.915 { 00:10:58.915 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:58.915 "dma_device_type": 2 00:10:58.915 }, 00:10:58.915 { 00:10:58.915 "dma_device_id": "system", 00:10:58.915 "dma_device_type": 1 00:10:58.915 }, 00:10:58.915 { 00:10:58.915 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:58.915 "dma_device_type": 2 00:10:58.915 } 00:10:58.915 ], 00:10:58.915 "driver_specific": { 00:10:58.915 "raid": { 00:10:58.915 "uuid": "01f22662-bc99-4826-948f-2f5f386471f9", 00:10:58.915 "strip_size_kb": 64, 00:10:58.915 "state": "online", 00:10:58.915 "raid_level": "raid0", 00:10:58.915 "superblock": true, 00:10:58.915 "num_base_bdevs": 4, 00:10:58.915 "num_base_bdevs_discovered": 4, 00:10:58.915 "num_base_bdevs_operational": 4, 00:10:58.915 "base_bdevs_list": [ 00:10:58.915 { 00:10:58.915 "name": "BaseBdev1", 00:10:58.915 "uuid": "fdc91caf-941a-4265-b629-eb67820313f9", 00:10:58.915 "is_configured": true, 00:10:58.915 "data_offset": 2048, 00:10:58.915 "data_size": 63488 00:10:58.915 }, 00:10:58.915 { 00:10:58.915 "name": "BaseBdev2", 00:10:58.915 "uuid": "34ba2329-9ba9-4605-8ef7-38847e6a5770", 00:10:58.915 "is_configured": true, 00:10:58.915 "data_offset": 2048, 00:10:58.915 "data_size": 63488 00:10:58.915 }, 00:10:58.915 { 00:10:58.915 "name": "BaseBdev3", 00:10:58.915 "uuid": "1f7ffec7-a57f-4071-adae-3e2894b82c8c", 00:10:58.915 "is_configured": true, 00:10:58.915 "data_offset": 2048, 00:10:58.915 "data_size": 63488 00:10:58.915 }, 00:10:58.915 { 00:10:58.915 "name": "BaseBdev4", 00:10:58.915 "uuid": "d373aed6-8b75-4d58-8347-b3d5a22f6fc6", 00:10:58.915 "is_configured": true, 00:10:58.915 "data_offset": 2048, 00:10:58.915 "data_size": 63488 00:10:58.915 } 00:10:58.915 ] 00:10:58.915 } 00:10:58.915 } 00:10:58.915 }' 00:10:58.915 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:58.915 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:10:58.915 BaseBdev2 00:10:58.915 BaseBdev3 00:10:58.915 BaseBdev4' 00:10:58.915 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:58.915 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:58.915 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:58.915 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:10:58.915 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:58.915 15:16:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.915 15:16:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:58.915 15:16:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.915 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:58.915 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:58.915 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:58.915 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:58.915 15:16:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.915 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:58.915 15:16:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:58.915 15:16:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.915 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:58.915 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:58.915 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:58.915 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:58.915 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:58.915 15:16:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.915 15:16:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:58.915 15:16:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:59.177 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:59.177 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:59.177 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:59.177 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:59.177 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:59.177 15:16:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:59.177 15:16:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:59.177 15:16:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:59.177 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:59.177 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:59.177 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:59.177 15:16:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:59.177 15:16:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:59.177 [2024-11-19 15:16:49.326647] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:59.177 [2024-11-19 15:16:49.326681] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:59.177 [2024-11-19 15:16:49.326733] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:59.177 15:16:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:59.177 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:10:59.177 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:10:59.177 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:59.177 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:10:59.177 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:10:59.177 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 3 00:10:59.177 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:59.177 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:10:59.177 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:59.177 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:59.177 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:59.177 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:59.177 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:59.177 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:59.177 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:59.177 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:59.177 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:59.177 15:16:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:59.177 15:16:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:59.177 15:16:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:59.177 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:59.177 "name": "Existed_Raid", 00:10:59.177 "uuid": "01f22662-bc99-4826-948f-2f5f386471f9", 00:10:59.177 "strip_size_kb": 64, 00:10:59.177 "state": "offline", 00:10:59.177 "raid_level": "raid0", 00:10:59.177 "superblock": true, 00:10:59.177 "num_base_bdevs": 4, 00:10:59.177 "num_base_bdevs_discovered": 3, 00:10:59.177 "num_base_bdevs_operational": 3, 00:10:59.177 "base_bdevs_list": [ 00:10:59.177 { 00:10:59.177 "name": null, 00:10:59.177 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:59.177 "is_configured": false, 00:10:59.177 "data_offset": 0, 00:10:59.177 "data_size": 63488 00:10:59.177 }, 00:10:59.177 { 00:10:59.177 "name": "BaseBdev2", 00:10:59.177 "uuid": "34ba2329-9ba9-4605-8ef7-38847e6a5770", 00:10:59.177 "is_configured": true, 00:10:59.177 "data_offset": 2048, 00:10:59.177 "data_size": 63488 00:10:59.177 }, 00:10:59.177 { 00:10:59.177 "name": "BaseBdev3", 00:10:59.177 "uuid": "1f7ffec7-a57f-4071-adae-3e2894b82c8c", 00:10:59.177 "is_configured": true, 00:10:59.177 "data_offset": 2048, 00:10:59.177 "data_size": 63488 00:10:59.177 }, 00:10:59.177 { 00:10:59.177 "name": "BaseBdev4", 00:10:59.177 "uuid": "d373aed6-8b75-4d58-8347-b3d5a22f6fc6", 00:10:59.177 "is_configured": true, 00:10:59.177 "data_offset": 2048, 00:10:59.177 "data_size": 63488 00:10:59.178 } 00:10:59.178 ] 00:10:59.178 }' 00:10:59.178 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:59.178 15:16:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:59.748 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:10:59.748 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:59.748 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:59.748 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:59.748 15:16:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:59.748 15:16:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:59.748 15:16:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:59.748 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:59.748 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:59.748 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:10:59.748 15:16:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:59.748 15:16:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:59.748 [2024-11-19 15:16:49.846891] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:59.748 15:16:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:59.748 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:59.748 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:59.748 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:59.748 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:59.748 15:16:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:59.748 15:16:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:59.748 15:16:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:59.748 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:59.748 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:59.748 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:10:59.748 15:16:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:59.748 15:16:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:59.748 [2024-11-19 15:16:49.923562] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:59.748 15:16:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:59.748 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:59.748 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:59.748 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:59.748 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:59.748 15:16:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:59.748 15:16:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:59.748 15:16:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:59.748 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:59.748 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:59.748 15:16:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:10:59.748 15:16:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:59.749 15:16:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:59.749 [2024-11-19 15:16:49.999242] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:10:59.749 [2024-11-19 15:16:49.999358] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:10:59.749 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:59.749 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:59.749 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:59.749 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:10:59.749 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:59.749 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:59.749 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:59.749 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:59.749 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:10:59.749 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:10:59.749 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:10:59.749 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:10:59.749 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:59.749 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:59.749 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:59.749 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:59.749 BaseBdev2 00:10:59.749 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:59.749 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:10:59.749 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:10:59.749 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:59.749 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:59.749 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:59.749 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:59.749 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:59.749 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:59.749 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:00.009 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.009 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:11:00.009 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.009 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:00.009 [ 00:11:00.009 { 00:11:00.009 "name": "BaseBdev2", 00:11:00.009 "aliases": [ 00:11:00.009 "8e663aa2-942d-4acf-8b5d-e89168083bed" 00:11:00.009 ], 00:11:00.009 "product_name": "Malloc disk", 00:11:00.009 "block_size": 512, 00:11:00.009 "num_blocks": 65536, 00:11:00.009 "uuid": "8e663aa2-942d-4acf-8b5d-e89168083bed", 00:11:00.009 "assigned_rate_limits": { 00:11:00.009 "rw_ios_per_sec": 0, 00:11:00.009 "rw_mbytes_per_sec": 0, 00:11:00.009 "r_mbytes_per_sec": 0, 00:11:00.009 "w_mbytes_per_sec": 0 00:11:00.009 }, 00:11:00.009 "claimed": false, 00:11:00.009 "zoned": false, 00:11:00.009 "supported_io_types": { 00:11:00.009 "read": true, 00:11:00.009 "write": true, 00:11:00.009 "unmap": true, 00:11:00.009 "flush": true, 00:11:00.009 "reset": true, 00:11:00.009 "nvme_admin": false, 00:11:00.009 "nvme_io": false, 00:11:00.009 "nvme_io_md": false, 00:11:00.009 "write_zeroes": true, 00:11:00.009 "zcopy": true, 00:11:00.009 "get_zone_info": false, 00:11:00.009 "zone_management": false, 00:11:00.009 "zone_append": false, 00:11:00.009 "compare": false, 00:11:00.009 "compare_and_write": false, 00:11:00.009 "abort": true, 00:11:00.009 "seek_hole": false, 00:11:00.009 "seek_data": false, 00:11:00.009 "copy": true, 00:11:00.009 "nvme_iov_md": false 00:11:00.009 }, 00:11:00.009 "memory_domains": [ 00:11:00.009 { 00:11:00.009 "dma_device_id": "system", 00:11:00.009 "dma_device_type": 1 00:11:00.009 }, 00:11:00.009 { 00:11:00.009 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:00.009 "dma_device_type": 2 00:11:00.010 } 00:11:00.010 ], 00:11:00.010 "driver_specific": {} 00:11:00.010 } 00:11:00.010 ] 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:00.010 BaseBdev3 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:00.010 [ 00:11:00.010 { 00:11:00.010 "name": "BaseBdev3", 00:11:00.010 "aliases": [ 00:11:00.010 "9af4da46-902c-4566-bab4-d5213957e08e" 00:11:00.010 ], 00:11:00.010 "product_name": "Malloc disk", 00:11:00.010 "block_size": 512, 00:11:00.010 "num_blocks": 65536, 00:11:00.010 "uuid": "9af4da46-902c-4566-bab4-d5213957e08e", 00:11:00.010 "assigned_rate_limits": { 00:11:00.010 "rw_ios_per_sec": 0, 00:11:00.010 "rw_mbytes_per_sec": 0, 00:11:00.010 "r_mbytes_per_sec": 0, 00:11:00.010 "w_mbytes_per_sec": 0 00:11:00.010 }, 00:11:00.010 "claimed": false, 00:11:00.010 "zoned": false, 00:11:00.010 "supported_io_types": { 00:11:00.010 "read": true, 00:11:00.010 "write": true, 00:11:00.010 "unmap": true, 00:11:00.010 "flush": true, 00:11:00.010 "reset": true, 00:11:00.010 "nvme_admin": false, 00:11:00.010 "nvme_io": false, 00:11:00.010 "nvme_io_md": false, 00:11:00.010 "write_zeroes": true, 00:11:00.010 "zcopy": true, 00:11:00.010 "get_zone_info": false, 00:11:00.010 "zone_management": false, 00:11:00.010 "zone_append": false, 00:11:00.010 "compare": false, 00:11:00.010 "compare_and_write": false, 00:11:00.010 "abort": true, 00:11:00.010 "seek_hole": false, 00:11:00.010 "seek_data": false, 00:11:00.010 "copy": true, 00:11:00.010 "nvme_iov_md": false 00:11:00.010 }, 00:11:00.010 "memory_domains": [ 00:11:00.010 { 00:11:00.010 "dma_device_id": "system", 00:11:00.010 "dma_device_type": 1 00:11:00.010 }, 00:11:00.010 { 00:11:00.010 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:00.010 "dma_device_type": 2 00:11:00.010 } 00:11:00.010 ], 00:11:00.010 "driver_specific": {} 00:11:00.010 } 00:11:00.010 ] 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:00.010 BaseBdev4 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:00.010 [ 00:11:00.010 { 00:11:00.010 "name": "BaseBdev4", 00:11:00.010 "aliases": [ 00:11:00.010 "c3d7c098-87b2-46e8-8334-e8eac47e34aa" 00:11:00.010 ], 00:11:00.010 "product_name": "Malloc disk", 00:11:00.010 "block_size": 512, 00:11:00.010 "num_blocks": 65536, 00:11:00.010 "uuid": "c3d7c098-87b2-46e8-8334-e8eac47e34aa", 00:11:00.010 "assigned_rate_limits": { 00:11:00.010 "rw_ios_per_sec": 0, 00:11:00.010 "rw_mbytes_per_sec": 0, 00:11:00.010 "r_mbytes_per_sec": 0, 00:11:00.010 "w_mbytes_per_sec": 0 00:11:00.010 }, 00:11:00.010 "claimed": false, 00:11:00.010 "zoned": false, 00:11:00.010 "supported_io_types": { 00:11:00.010 "read": true, 00:11:00.010 "write": true, 00:11:00.010 "unmap": true, 00:11:00.010 "flush": true, 00:11:00.010 "reset": true, 00:11:00.010 "nvme_admin": false, 00:11:00.010 "nvme_io": false, 00:11:00.010 "nvme_io_md": false, 00:11:00.010 "write_zeroes": true, 00:11:00.010 "zcopy": true, 00:11:00.010 "get_zone_info": false, 00:11:00.010 "zone_management": false, 00:11:00.010 "zone_append": false, 00:11:00.010 "compare": false, 00:11:00.010 "compare_and_write": false, 00:11:00.010 "abort": true, 00:11:00.010 "seek_hole": false, 00:11:00.010 "seek_data": false, 00:11:00.010 "copy": true, 00:11:00.010 "nvme_iov_md": false 00:11:00.010 }, 00:11:00.010 "memory_domains": [ 00:11:00.010 { 00:11:00.010 "dma_device_id": "system", 00:11:00.010 "dma_device_type": 1 00:11:00.010 }, 00:11:00.010 { 00:11:00.010 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:00.010 "dma_device_type": 2 00:11:00.010 } 00:11:00.010 ], 00:11:00.010 "driver_specific": {} 00:11:00.010 } 00:11:00.010 ] 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:00.010 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:00.011 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.011 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:00.011 [2024-11-19 15:16:50.246678] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:11:00.011 [2024-11-19 15:16:50.246804] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:11:00.011 [2024-11-19 15:16:50.246867] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:00.011 [2024-11-19 15:16:50.249024] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:00.011 [2024-11-19 15:16:50.249113] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:00.011 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.011 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:11:00.011 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:00.011 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:00.011 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:11:00.011 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:00.011 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:00.011 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:00.011 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:00.011 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:00.011 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:00.011 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:00.011 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:00.011 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.011 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:00.011 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.011 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:00.011 "name": "Existed_Raid", 00:11:00.011 "uuid": "1e7e5968-fde5-402f-8313-91ca5176d8cc", 00:11:00.011 "strip_size_kb": 64, 00:11:00.011 "state": "configuring", 00:11:00.011 "raid_level": "raid0", 00:11:00.011 "superblock": true, 00:11:00.011 "num_base_bdevs": 4, 00:11:00.011 "num_base_bdevs_discovered": 3, 00:11:00.011 "num_base_bdevs_operational": 4, 00:11:00.011 "base_bdevs_list": [ 00:11:00.011 { 00:11:00.011 "name": "BaseBdev1", 00:11:00.011 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:00.011 "is_configured": false, 00:11:00.011 "data_offset": 0, 00:11:00.011 "data_size": 0 00:11:00.011 }, 00:11:00.011 { 00:11:00.011 "name": "BaseBdev2", 00:11:00.011 "uuid": "8e663aa2-942d-4acf-8b5d-e89168083bed", 00:11:00.011 "is_configured": true, 00:11:00.011 "data_offset": 2048, 00:11:00.011 "data_size": 63488 00:11:00.011 }, 00:11:00.011 { 00:11:00.011 "name": "BaseBdev3", 00:11:00.011 "uuid": "9af4da46-902c-4566-bab4-d5213957e08e", 00:11:00.011 "is_configured": true, 00:11:00.011 "data_offset": 2048, 00:11:00.011 "data_size": 63488 00:11:00.011 }, 00:11:00.011 { 00:11:00.011 "name": "BaseBdev4", 00:11:00.011 "uuid": "c3d7c098-87b2-46e8-8334-e8eac47e34aa", 00:11:00.011 "is_configured": true, 00:11:00.011 "data_offset": 2048, 00:11:00.011 "data_size": 63488 00:11:00.011 } 00:11:00.011 ] 00:11:00.011 }' 00:11:00.011 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:00.011 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:00.581 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:11:00.581 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.581 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:00.581 [2024-11-19 15:16:50.661930] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:11:00.581 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.581 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:11:00.581 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:00.581 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:00.581 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:11:00.581 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:00.581 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:00.581 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:00.581 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:00.581 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:00.581 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:00.581 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:00.581 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:00.581 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.581 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:00.581 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.581 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:00.581 "name": "Existed_Raid", 00:11:00.581 "uuid": "1e7e5968-fde5-402f-8313-91ca5176d8cc", 00:11:00.581 "strip_size_kb": 64, 00:11:00.581 "state": "configuring", 00:11:00.581 "raid_level": "raid0", 00:11:00.581 "superblock": true, 00:11:00.581 "num_base_bdevs": 4, 00:11:00.581 "num_base_bdevs_discovered": 2, 00:11:00.581 "num_base_bdevs_operational": 4, 00:11:00.581 "base_bdevs_list": [ 00:11:00.581 { 00:11:00.581 "name": "BaseBdev1", 00:11:00.581 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:00.581 "is_configured": false, 00:11:00.581 "data_offset": 0, 00:11:00.581 "data_size": 0 00:11:00.581 }, 00:11:00.581 { 00:11:00.581 "name": null, 00:11:00.581 "uuid": "8e663aa2-942d-4acf-8b5d-e89168083bed", 00:11:00.581 "is_configured": false, 00:11:00.581 "data_offset": 0, 00:11:00.581 "data_size": 63488 00:11:00.581 }, 00:11:00.581 { 00:11:00.581 "name": "BaseBdev3", 00:11:00.581 "uuid": "9af4da46-902c-4566-bab4-d5213957e08e", 00:11:00.581 "is_configured": true, 00:11:00.581 "data_offset": 2048, 00:11:00.581 "data_size": 63488 00:11:00.581 }, 00:11:00.581 { 00:11:00.581 "name": "BaseBdev4", 00:11:00.581 "uuid": "c3d7c098-87b2-46e8-8334-e8eac47e34aa", 00:11:00.581 "is_configured": true, 00:11:00.581 "data_offset": 2048, 00:11:00.581 "data_size": 63488 00:11:00.581 } 00:11:00.581 ] 00:11:00.581 }' 00:11:00.581 15:16:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:00.581 15:16:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:00.841 15:16:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:00.841 15:16:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:11:00.841 15:16:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.841 15:16:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:00.841 15:16:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.841 15:16:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:11:00.841 15:16:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:11:00.841 15:16:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.841 15:16:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:00.841 [2024-11-19 15:16:51.177777] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:00.841 BaseBdev1 00:11:01.101 15:16:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:01.101 15:16:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:11:01.101 15:16:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:11:01.101 15:16:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:01.101 15:16:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:11:01.101 15:16:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:01.101 15:16:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:01.101 15:16:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:01.101 15:16:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:01.101 15:16:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:01.101 15:16:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:01.101 15:16:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:11:01.101 15:16:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:01.101 15:16:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:01.101 [ 00:11:01.101 { 00:11:01.101 "name": "BaseBdev1", 00:11:01.101 "aliases": [ 00:11:01.101 "8de663ab-dd1a-4db9-9581-c3e81d1a967c" 00:11:01.101 ], 00:11:01.101 "product_name": "Malloc disk", 00:11:01.101 "block_size": 512, 00:11:01.101 "num_blocks": 65536, 00:11:01.101 "uuid": "8de663ab-dd1a-4db9-9581-c3e81d1a967c", 00:11:01.101 "assigned_rate_limits": { 00:11:01.101 "rw_ios_per_sec": 0, 00:11:01.101 "rw_mbytes_per_sec": 0, 00:11:01.101 "r_mbytes_per_sec": 0, 00:11:01.101 "w_mbytes_per_sec": 0 00:11:01.101 }, 00:11:01.101 "claimed": true, 00:11:01.101 "claim_type": "exclusive_write", 00:11:01.101 "zoned": false, 00:11:01.101 "supported_io_types": { 00:11:01.101 "read": true, 00:11:01.101 "write": true, 00:11:01.101 "unmap": true, 00:11:01.101 "flush": true, 00:11:01.101 "reset": true, 00:11:01.101 "nvme_admin": false, 00:11:01.101 "nvme_io": false, 00:11:01.101 "nvme_io_md": false, 00:11:01.101 "write_zeroes": true, 00:11:01.101 "zcopy": true, 00:11:01.101 "get_zone_info": false, 00:11:01.101 "zone_management": false, 00:11:01.101 "zone_append": false, 00:11:01.101 "compare": false, 00:11:01.101 "compare_and_write": false, 00:11:01.101 "abort": true, 00:11:01.101 "seek_hole": false, 00:11:01.101 "seek_data": false, 00:11:01.101 "copy": true, 00:11:01.101 "nvme_iov_md": false 00:11:01.101 }, 00:11:01.101 "memory_domains": [ 00:11:01.101 { 00:11:01.101 "dma_device_id": "system", 00:11:01.101 "dma_device_type": 1 00:11:01.101 }, 00:11:01.101 { 00:11:01.101 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:01.101 "dma_device_type": 2 00:11:01.101 } 00:11:01.101 ], 00:11:01.101 "driver_specific": {} 00:11:01.101 } 00:11:01.101 ] 00:11:01.101 15:16:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:01.101 15:16:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:11:01.101 15:16:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:11:01.101 15:16:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:01.101 15:16:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:01.101 15:16:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:11:01.101 15:16:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:01.101 15:16:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:01.101 15:16:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:01.101 15:16:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:01.101 15:16:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:01.101 15:16:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:01.101 15:16:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:01.101 15:16:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:01.101 15:16:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:01.101 15:16:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:01.101 15:16:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:01.101 15:16:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:01.101 "name": "Existed_Raid", 00:11:01.101 "uuid": "1e7e5968-fde5-402f-8313-91ca5176d8cc", 00:11:01.101 "strip_size_kb": 64, 00:11:01.101 "state": "configuring", 00:11:01.101 "raid_level": "raid0", 00:11:01.101 "superblock": true, 00:11:01.101 "num_base_bdevs": 4, 00:11:01.101 "num_base_bdevs_discovered": 3, 00:11:01.101 "num_base_bdevs_operational": 4, 00:11:01.101 "base_bdevs_list": [ 00:11:01.101 { 00:11:01.101 "name": "BaseBdev1", 00:11:01.101 "uuid": "8de663ab-dd1a-4db9-9581-c3e81d1a967c", 00:11:01.101 "is_configured": true, 00:11:01.101 "data_offset": 2048, 00:11:01.101 "data_size": 63488 00:11:01.101 }, 00:11:01.101 { 00:11:01.101 "name": null, 00:11:01.101 "uuid": "8e663aa2-942d-4acf-8b5d-e89168083bed", 00:11:01.101 "is_configured": false, 00:11:01.101 "data_offset": 0, 00:11:01.101 "data_size": 63488 00:11:01.101 }, 00:11:01.101 { 00:11:01.101 "name": "BaseBdev3", 00:11:01.101 "uuid": "9af4da46-902c-4566-bab4-d5213957e08e", 00:11:01.101 "is_configured": true, 00:11:01.101 "data_offset": 2048, 00:11:01.101 "data_size": 63488 00:11:01.101 }, 00:11:01.101 { 00:11:01.101 "name": "BaseBdev4", 00:11:01.101 "uuid": "c3d7c098-87b2-46e8-8334-e8eac47e34aa", 00:11:01.101 "is_configured": true, 00:11:01.101 "data_offset": 2048, 00:11:01.101 "data_size": 63488 00:11:01.101 } 00:11:01.101 ] 00:11:01.101 }' 00:11:01.101 15:16:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:01.101 15:16:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:01.361 15:16:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:01.361 15:16:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:01.361 15:16:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:01.361 15:16:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:11:01.361 15:16:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:01.361 15:16:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:11:01.361 15:16:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:11:01.361 15:16:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:01.361 15:16:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:01.621 [2024-11-19 15:16:51.700939] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:11:01.621 15:16:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:01.621 15:16:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:11:01.621 15:16:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:01.621 15:16:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:01.621 15:16:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:11:01.621 15:16:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:01.621 15:16:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:01.621 15:16:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:01.621 15:16:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:01.621 15:16:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:01.621 15:16:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:01.621 15:16:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:01.621 15:16:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:01.621 15:16:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:01.621 15:16:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:01.621 15:16:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:01.621 15:16:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:01.621 "name": "Existed_Raid", 00:11:01.621 "uuid": "1e7e5968-fde5-402f-8313-91ca5176d8cc", 00:11:01.621 "strip_size_kb": 64, 00:11:01.621 "state": "configuring", 00:11:01.621 "raid_level": "raid0", 00:11:01.621 "superblock": true, 00:11:01.621 "num_base_bdevs": 4, 00:11:01.621 "num_base_bdevs_discovered": 2, 00:11:01.621 "num_base_bdevs_operational": 4, 00:11:01.621 "base_bdevs_list": [ 00:11:01.621 { 00:11:01.621 "name": "BaseBdev1", 00:11:01.621 "uuid": "8de663ab-dd1a-4db9-9581-c3e81d1a967c", 00:11:01.621 "is_configured": true, 00:11:01.621 "data_offset": 2048, 00:11:01.621 "data_size": 63488 00:11:01.621 }, 00:11:01.621 { 00:11:01.621 "name": null, 00:11:01.621 "uuid": "8e663aa2-942d-4acf-8b5d-e89168083bed", 00:11:01.621 "is_configured": false, 00:11:01.621 "data_offset": 0, 00:11:01.621 "data_size": 63488 00:11:01.621 }, 00:11:01.621 { 00:11:01.621 "name": null, 00:11:01.621 "uuid": "9af4da46-902c-4566-bab4-d5213957e08e", 00:11:01.621 "is_configured": false, 00:11:01.621 "data_offset": 0, 00:11:01.621 "data_size": 63488 00:11:01.621 }, 00:11:01.621 { 00:11:01.621 "name": "BaseBdev4", 00:11:01.621 "uuid": "c3d7c098-87b2-46e8-8334-e8eac47e34aa", 00:11:01.621 "is_configured": true, 00:11:01.621 "data_offset": 2048, 00:11:01.621 "data_size": 63488 00:11:01.621 } 00:11:01.621 ] 00:11:01.621 }' 00:11:01.621 15:16:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:01.621 15:16:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:01.881 15:16:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:01.882 15:16:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:01.882 15:16:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:11:01.882 15:16:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:01.882 15:16:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:01.882 15:16:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:11:01.882 15:16:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:11:01.882 15:16:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:01.882 15:16:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:01.882 [2024-11-19 15:16:52.188124] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:01.882 15:16:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:01.882 15:16:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:11:01.882 15:16:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:01.882 15:16:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:01.882 15:16:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:11:01.882 15:16:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:01.882 15:16:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:01.882 15:16:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:01.882 15:16:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:01.882 15:16:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:01.882 15:16:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:01.882 15:16:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:01.882 15:16:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:01.882 15:16:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:01.882 15:16:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:02.142 15:16:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:02.142 15:16:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:02.142 "name": "Existed_Raid", 00:11:02.142 "uuid": "1e7e5968-fde5-402f-8313-91ca5176d8cc", 00:11:02.142 "strip_size_kb": 64, 00:11:02.142 "state": "configuring", 00:11:02.142 "raid_level": "raid0", 00:11:02.142 "superblock": true, 00:11:02.142 "num_base_bdevs": 4, 00:11:02.142 "num_base_bdevs_discovered": 3, 00:11:02.142 "num_base_bdevs_operational": 4, 00:11:02.142 "base_bdevs_list": [ 00:11:02.142 { 00:11:02.142 "name": "BaseBdev1", 00:11:02.142 "uuid": "8de663ab-dd1a-4db9-9581-c3e81d1a967c", 00:11:02.142 "is_configured": true, 00:11:02.142 "data_offset": 2048, 00:11:02.142 "data_size": 63488 00:11:02.142 }, 00:11:02.142 { 00:11:02.142 "name": null, 00:11:02.142 "uuid": "8e663aa2-942d-4acf-8b5d-e89168083bed", 00:11:02.142 "is_configured": false, 00:11:02.142 "data_offset": 0, 00:11:02.142 "data_size": 63488 00:11:02.142 }, 00:11:02.142 { 00:11:02.142 "name": "BaseBdev3", 00:11:02.142 "uuid": "9af4da46-902c-4566-bab4-d5213957e08e", 00:11:02.142 "is_configured": true, 00:11:02.142 "data_offset": 2048, 00:11:02.142 "data_size": 63488 00:11:02.142 }, 00:11:02.142 { 00:11:02.142 "name": "BaseBdev4", 00:11:02.142 "uuid": "c3d7c098-87b2-46e8-8334-e8eac47e34aa", 00:11:02.142 "is_configured": true, 00:11:02.142 "data_offset": 2048, 00:11:02.142 "data_size": 63488 00:11:02.142 } 00:11:02.142 ] 00:11:02.142 }' 00:11:02.143 15:16:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:02.143 15:16:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:02.403 15:16:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:02.403 15:16:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:02.403 15:16:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:02.403 15:16:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:11:02.403 15:16:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:02.403 15:16:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:11:02.403 15:16:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:11:02.403 15:16:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:02.403 15:16:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:02.403 [2024-11-19 15:16:52.675411] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:02.403 15:16:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:02.403 15:16:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:11:02.403 15:16:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:02.403 15:16:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:02.403 15:16:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:11:02.403 15:16:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:02.403 15:16:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:02.403 15:16:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:02.403 15:16:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:02.403 15:16:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:02.403 15:16:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:02.403 15:16:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:02.403 15:16:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:02.403 15:16:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:02.403 15:16:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:02.403 15:16:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:02.663 15:16:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:02.663 "name": "Existed_Raid", 00:11:02.663 "uuid": "1e7e5968-fde5-402f-8313-91ca5176d8cc", 00:11:02.663 "strip_size_kb": 64, 00:11:02.663 "state": "configuring", 00:11:02.663 "raid_level": "raid0", 00:11:02.663 "superblock": true, 00:11:02.663 "num_base_bdevs": 4, 00:11:02.663 "num_base_bdevs_discovered": 2, 00:11:02.663 "num_base_bdevs_operational": 4, 00:11:02.663 "base_bdevs_list": [ 00:11:02.663 { 00:11:02.663 "name": null, 00:11:02.663 "uuid": "8de663ab-dd1a-4db9-9581-c3e81d1a967c", 00:11:02.663 "is_configured": false, 00:11:02.663 "data_offset": 0, 00:11:02.663 "data_size": 63488 00:11:02.663 }, 00:11:02.663 { 00:11:02.663 "name": null, 00:11:02.663 "uuid": "8e663aa2-942d-4acf-8b5d-e89168083bed", 00:11:02.663 "is_configured": false, 00:11:02.663 "data_offset": 0, 00:11:02.663 "data_size": 63488 00:11:02.663 }, 00:11:02.663 { 00:11:02.663 "name": "BaseBdev3", 00:11:02.663 "uuid": "9af4da46-902c-4566-bab4-d5213957e08e", 00:11:02.663 "is_configured": true, 00:11:02.663 "data_offset": 2048, 00:11:02.663 "data_size": 63488 00:11:02.663 }, 00:11:02.663 { 00:11:02.663 "name": "BaseBdev4", 00:11:02.663 "uuid": "c3d7c098-87b2-46e8-8334-e8eac47e34aa", 00:11:02.663 "is_configured": true, 00:11:02.663 "data_offset": 2048, 00:11:02.663 "data_size": 63488 00:11:02.663 } 00:11:02.663 ] 00:11:02.663 }' 00:11:02.663 15:16:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:02.663 15:16:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:02.923 15:16:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:11:02.923 15:16:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:02.923 15:16:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:02.923 15:16:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:02.923 15:16:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:02.923 15:16:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:11:02.923 15:16:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:11:02.923 15:16:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:02.923 15:16:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:02.923 [2024-11-19 15:16:53.154488] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:02.923 15:16:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:02.923 15:16:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:11:02.923 15:16:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:02.923 15:16:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:02.923 15:16:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:11:02.923 15:16:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:02.923 15:16:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:02.923 15:16:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:02.923 15:16:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:02.923 15:16:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:02.923 15:16:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:02.923 15:16:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:02.923 15:16:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:02.923 15:16:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:02.923 15:16:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:02.923 15:16:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:02.923 15:16:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:02.923 "name": "Existed_Raid", 00:11:02.923 "uuid": "1e7e5968-fde5-402f-8313-91ca5176d8cc", 00:11:02.923 "strip_size_kb": 64, 00:11:02.923 "state": "configuring", 00:11:02.923 "raid_level": "raid0", 00:11:02.923 "superblock": true, 00:11:02.923 "num_base_bdevs": 4, 00:11:02.923 "num_base_bdevs_discovered": 3, 00:11:02.923 "num_base_bdevs_operational": 4, 00:11:02.923 "base_bdevs_list": [ 00:11:02.923 { 00:11:02.923 "name": null, 00:11:02.923 "uuid": "8de663ab-dd1a-4db9-9581-c3e81d1a967c", 00:11:02.923 "is_configured": false, 00:11:02.923 "data_offset": 0, 00:11:02.923 "data_size": 63488 00:11:02.923 }, 00:11:02.923 { 00:11:02.923 "name": "BaseBdev2", 00:11:02.923 "uuid": "8e663aa2-942d-4acf-8b5d-e89168083bed", 00:11:02.923 "is_configured": true, 00:11:02.923 "data_offset": 2048, 00:11:02.923 "data_size": 63488 00:11:02.923 }, 00:11:02.923 { 00:11:02.923 "name": "BaseBdev3", 00:11:02.923 "uuid": "9af4da46-902c-4566-bab4-d5213957e08e", 00:11:02.923 "is_configured": true, 00:11:02.923 "data_offset": 2048, 00:11:02.923 "data_size": 63488 00:11:02.923 }, 00:11:02.923 { 00:11:02.923 "name": "BaseBdev4", 00:11:02.923 "uuid": "c3d7c098-87b2-46e8-8334-e8eac47e34aa", 00:11:02.923 "is_configured": true, 00:11:02.923 "data_offset": 2048, 00:11:02.923 "data_size": 63488 00:11:02.923 } 00:11:02.923 ] 00:11:02.923 }' 00:11:02.923 15:16:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:02.923 15:16:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:03.493 15:16:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:11:03.493 15:16:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:03.493 15:16:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.493 15:16:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:03.493 15:16:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.493 15:16:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:11:03.493 15:16:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:03.493 15:16:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.493 15:16:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:03.493 15:16:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:11:03.493 15:16:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.493 15:16:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 8de663ab-dd1a-4db9-9581-c3e81d1a967c 00:11:03.493 15:16:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.493 15:16:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:03.493 [2024-11-19 15:16:53.662707] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:11:03.493 [2024-11-19 15:16:53.663056] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:11:03.493 [2024-11-19 15:16:53.663108] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:11:03.493 [2024-11-19 15:16:53.663415] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:11:03.493 [2024-11-19 15:16:53.663571] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:11:03.493 [2024-11-19 15:16:53.663612] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:11:03.493 NewBaseBdev 00:11:03.493 [2024-11-19 15:16:53.663782] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:03.493 15:16:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.494 15:16:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:11:03.494 15:16:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:11:03.494 15:16:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:03.494 15:16:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:11:03.494 15:16:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:03.494 15:16:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:03.494 15:16:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:03.494 15:16:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.494 15:16:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:03.494 15:16:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.494 15:16:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:11:03.494 15:16:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.494 15:16:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:03.494 [ 00:11:03.494 { 00:11:03.494 "name": "NewBaseBdev", 00:11:03.494 "aliases": [ 00:11:03.494 "8de663ab-dd1a-4db9-9581-c3e81d1a967c" 00:11:03.494 ], 00:11:03.494 "product_name": "Malloc disk", 00:11:03.494 "block_size": 512, 00:11:03.494 "num_blocks": 65536, 00:11:03.494 "uuid": "8de663ab-dd1a-4db9-9581-c3e81d1a967c", 00:11:03.494 "assigned_rate_limits": { 00:11:03.494 "rw_ios_per_sec": 0, 00:11:03.494 "rw_mbytes_per_sec": 0, 00:11:03.494 "r_mbytes_per_sec": 0, 00:11:03.494 "w_mbytes_per_sec": 0 00:11:03.494 }, 00:11:03.494 "claimed": true, 00:11:03.494 "claim_type": "exclusive_write", 00:11:03.494 "zoned": false, 00:11:03.494 "supported_io_types": { 00:11:03.494 "read": true, 00:11:03.494 "write": true, 00:11:03.494 "unmap": true, 00:11:03.494 "flush": true, 00:11:03.494 "reset": true, 00:11:03.494 "nvme_admin": false, 00:11:03.494 "nvme_io": false, 00:11:03.494 "nvme_io_md": false, 00:11:03.494 "write_zeroes": true, 00:11:03.494 "zcopy": true, 00:11:03.494 "get_zone_info": false, 00:11:03.494 "zone_management": false, 00:11:03.494 "zone_append": false, 00:11:03.494 "compare": false, 00:11:03.494 "compare_and_write": false, 00:11:03.494 "abort": true, 00:11:03.494 "seek_hole": false, 00:11:03.494 "seek_data": false, 00:11:03.494 "copy": true, 00:11:03.494 "nvme_iov_md": false 00:11:03.494 }, 00:11:03.494 "memory_domains": [ 00:11:03.494 { 00:11:03.494 "dma_device_id": "system", 00:11:03.494 "dma_device_type": 1 00:11:03.494 }, 00:11:03.494 { 00:11:03.494 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:03.494 "dma_device_type": 2 00:11:03.494 } 00:11:03.494 ], 00:11:03.494 "driver_specific": {} 00:11:03.494 } 00:11:03.494 ] 00:11:03.494 15:16:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.494 15:16:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:11:03.494 15:16:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:11:03.494 15:16:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:03.494 15:16:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:03.494 15:16:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:11:03.494 15:16:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:03.494 15:16:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:03.494 15:16:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:03.494 15:16:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:03.494 15:16:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:03.494 15:16:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:03.494 15:16:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:03.494 15:16:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:03.494 15:16:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.494 15:16:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:03.494 15:16:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.494 15:16:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:03.494 "name": "Existed_Raid", 00:11:03.494 "uuid": "1e7e5968-fde5-402f-8313-91ca5176d8cc", 00:11:03.494 "strip_size_kb": 64, 00:11:03.494 "state": "online", 00:11:03.494 "raid_level": "raid0", 00:11:03.494 "superblock": true, 00:11:03.494 "num_base_bdevs": 4, 00:11:03.494 "num_base_bdevs_discovered": 4, 00:11:03.494 "num_base_bdevs_operational": 4, 00:11:03.494 "base_bdevs_list": [ 00:11:03.494 { 00:11:03.494 "name": "NewBaseBdev", 00:11:03.494 "uuid": "8de663ab-dd1a-4db9-9581-c3e81d1a967c", 00:11:03.494 "is_configured": true, 00:11:03.494 "data_offset": 2048, 00:11:03.494 "data_size": 63488 00:11:03.494 }, 00:11:03.494 { 00:11:03.494 "name": "BaseBdev2", 00:11:03.494 "uuid": "8e663aa2-942d-4acf-8b5d-e89168083bed", 00:11:03.494 "is_configured": true, 00:11:03.494 "data_offset": 2048, 00:11:03.494 "data_size": 63488 00:11:03.494 }, 00:11:03.494 { 00:11:03.494 "name": "BaseBdev3", 00:11:03.494 "uuid": "9af4da46-902c-4566-bab4-d5213957e08e", 00:11:03.494 "is_configured": true, 00:11:03.494 "data_offset": 2048, 00:11:03.494 "data_size": 63488 00:11:03.494 }, 00:11:03.494 { 00:11:03.494 "name": "BaseBdev4", 00:11:03.494 "uuid": "c3d7c098-87b2-46e8-8334-e8eac47e34aa", 00:11:03.494 "is_configured": true, 00:11:03.494 "data_offset": 2048, 00:11:03.494 "data_size": 63488 00:11:03.494 } 00:11:03.494 ] 00:11:03.494 }' 00:11:03.494 15:16:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:03.494 15:16:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:04.065 15:16:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:11:04.065 15:16:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:11:04.065 15:16:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:11:04.065 15:16:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:11:04.065 15:16:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:11:04.065 15:16:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:11:04.065 15:16:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:11:04.065 15:16:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:11:04.065 15:16:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:04.065 15:16:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:04.065 [2024-11-19 15:16:54.198242] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:04.065 15:16:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:04.065 15:16:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:11:04.065 "name": "Existed_Raid", 00:11:04.065 "aliases": [ 00:11:04.065 "1e7e5968-fde5-402f-8313-91ca5176d8cc" 00:11:04.065 ], 00:11:04.065 "product_name": "Raid Volume", 00:11:04.065 "block_size": 512, 00:11:04.065 "num_blocks": 253952, 00:11:04.065 "uuid": "1e7e5968-fde5-402f-8313-91ca5176d8cc", 00:11:04.065 "assigned_rate_limits": { 00:11:04.065 "rw_ios_per_sec": 0, 00:11:04.065 "rw_mbytes_per_sec": 0, 00:11:04.065 "r_mbytes_per_sec": 0, 00:11:04.065 "w_mbytes_per_sec": 0 00:11:04.065 }, 00:11:04.065 "claimed": false, 00:11:04.065 "zoned": false, 00:11:04.065 "supported_io_types": { 00:11:04.065 "read": true, 00:11:04.065 "write": true, 00:11:04.065 "unmap": true, 00:11:04.065 "flush": true, 00:11:04.065 "reset": true, 00:11:04.065 "nvme_admin": false, 00:11:04.065 "nvme_io": false, 00:11:04.065 "nvme_io_md": false, 00:11:04.065 "write_zeroes": true, 00:11:04.065 "zcopy": false, 00:11:04.065 "get_zone_info": false, 00:11:04.065 "zone_management": false, 00:11:04.065 "zone_append": false, 00:11:04.065 "compare": false, 00:11:04.065 "compare_and_write": false, 00:11:04.065 "abort": false, 00:11:04.065 "seek_hole": false, 00:11:04.065 "seek_data": false, 00:11:04.065 "copy": false, 00:11:04.065 "nvme_iov_md": false 00:11:04.065 }, 00:11:04.065 "memory_domains": [ 00:11:04.065 { 00:11:04.065 "dma_device_id": "system", 00:11:04.065 "dma_device_type": 1 00:11:04.065 }, 00:11:04.065 { 00:11:04.065 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:04.065 "dma_device_type": 2 00:11:04.065 }, 00:11:04.065 { 00:11:04.065 "dma_device_id": "system", 00:11:04.065 "dma_device_type": 1 00:11:04.065 }, 00:11:04.065 { 00:11:04.065 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:04.065 "dma_device_type": 2 00:11:04.065 }, 00:11:04.065 { 00:11:04.065 "dma_device_id": "system", 00:11:04.065 "dma_device_type": 1 00:11:04.065 }, 00:11:04.065 { 00:11:04.065 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:04.065 "dma_device_type": 2 00:11:04.065 }, 00:11:04.065 { 00:11:04.065 "dma_device_id": "system", 00:11:04.065 "dma_device_type": 1 00:11:04.065 }, 00:11:04.065 { 00:11:04.065 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:04.065 "dma_device_type": 2 00:11:04.065 } 00:11:04.065 ], 00:11:04.065 "driver_specific": { 00:11:04.065 "raid": { 00:11:04.065 "uuid": "1e7e5968-fde5-402f-8313-91ca5176d8cc", 00:11:04.065 "strip_size_kb": 64, 00:11:04.065 "state": "online", 00:11:04.065 "raid_level": "raid0", 00:11:04.065 "superblock": true, 00:11:04.065 "num_base_bdevs": 4, 00:11:04.065 "num_base_bdevs_discovered": 4, 00:11:04.065 "num_base_bdevs_operational": 4, 00:11:04.065 "base_bdevs_list": [ 00:11:04.065 { 00:11:04.065 "name": "NewBaseBdev", 00:11:04.065 "uuid": "8de663ab-dd1a-4db9-9581-c3e81d1a967c", 00:11:04.065 "is_configured": true, 00:11:04.065 "data_offset": 2048, 00:11:04.065 "data_size": 63488 00:11:04.065 }, 00:11:04.065 { 00:11:04.065 "name": "BaseBdev2", 00:11:04.065 "uuid": "8e663aa2-942d-4acf-8b5d-e89168083bed", 00:11:04.065 "is_configured": true, 00:11:04.065 "data_offset": 2048, 00:11:04.065 "data_size": 63488 00:11:04.065 }, 00:11:04.065 { 00:11:04.065 "name": "BaseBdev3", 00:11:04.065 "uuid": "9af4da46-902c-4566-bab4-d5213957e08e", 00:11:04.065 "is_configured": true, 00:11:04.065 "data_offset": 2048, 00:11:04.065 "data_size": 63488 00:11:04.065 }, 00:11:04.065 { 00:11:04.065 "name": "BaseBdev4", 00:11:04.065 "uuid": "c3d7c098-87b2-46e8-8334-e8eac47e34aa", 00:11:04.065 "is_configured": true, 00:11:04.065 "data_offset": 2048, 00:11:04.065 "data_size": 63488 00:11:04.065 } 00:11:04.065 ] 00:11:04.065 } 00:11:04.065 } 00:11:04.065 }' 00:11:04.065 15:16:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:11:04.065 15:16:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:11:04.065 BaseBdev2 00:11:04.065 BaseBdev3 00:11:04.065 BaseBdev4' 00:11:04.065 15:16:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:04.065 15:16:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:11:04.065 15:16:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:04.065 15:16:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:11:04.065 15:16:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:04.065 15:16:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:04.065 15:16:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:04.065 15:16:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:04.065 15:16:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:04.065 15:16:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:04.065 15:16:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:04.065 15:16:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:11:04.065 15:16:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:04.065 15:16:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:04.065 15:16:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:04.065 15:16:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:04.065 15:16:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:04.065 15:16:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:04.065 15:16:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:04.065 15:16:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:04.065 15:16:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:11:04.065 15:16:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:04.065 15:16:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:04.326 15:16:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:04.326 15:16:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:04.326 15:16:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:04.326 15:16:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:04.326 15:16:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:11:04.326 15:16:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:04.326 15:16:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:04.326 15:16:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:04.326 15:16:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:04.326 15:16:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:04.326 15:16:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:04.326 15:16:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:11:04.326 15:16:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:04.326 15:16:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:04.326 [2024-11-19 15:16:54.477331] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:11:04.326 [2024-11-19 15:16:54.477366] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:04.326 [2024-11-19 15:16:54.477449] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:04.326 [2024-11-19 15:16:54.477523] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:04.326 [2024-11-19 15:16:54.477534] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:11:04.326 15:16:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:04.326 15:16:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 81016 00:11:04.326 15:16:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 81016 ']' 00:11:04.326 15:16:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 81016 00:11:04.326 15:16:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:11:04.326 15:16:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:11:04.326 15:16:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 81016 00:11:04.326 15:16:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:11:04.326 15:16:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:11:04.327 15:16:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 81016' 00:11:04.327 killing process with pid 81016 00:11:04.327 15:16:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 81016 00:11:04.327 [2024-11-19 15:16:54.516382] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:04.327 15:16:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 81016 00:11:04.327 [2024-11-19 15:16:54.590868] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:04.587 15:16:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:11:04.587 00:11:04.587 real 0m9.797s 00:11:04.587 user 0m16.548s 00:11:04.587 sys 0m2.054s 00:11:04.587 15:16:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:04.587 ************************************ 00:11:04.587 END TEST raid_state_function_test_sb 00:11:04.587 ************************************ 00:11:04.587 15:16:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:04.847 15:16:54 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid0 4 00:11:04.847 15:16:54 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:11:04.847 15:16:54 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:04.847 15:16:54 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:04.847 ************************************ 00:11:04.847 START TEST raid_superblock_test 00:11:04.847 ************************************ 00:11:04.847 15:16:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid0 4 00:11:04.847 15:16:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid0 00:11:04.847 15:16:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:11:04.847 15:16:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:11:04.847 15:16:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:11:04.847 15:16:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:11:04.847 15:16:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:11:04.847 15:16:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:11:04.847 15:16:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:11:04.847 15:16:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:11:04.847 15:16:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:11:04.847 15:16:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:11:04.847 15:16:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:11:04.847 15:16:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:11:04.847 15:16:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid0 '!=' raid1 ']' 00:11:04.847 15:16:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:11:04.847 15:16:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:11:04.847 15:16:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:11:04.847 15:16:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=81670 00:11:04.847 15:16:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 81670 00:11:04.847 15:16:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 81670 ']' 00:11:04.847 15:16:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:04.847 15:16:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:04.847 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:04.847 15:16:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:04.847 15:16:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:04.847 15:16:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.847 [2024-11-19 15:16:55.049838] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:11:04.847 [2024-11-19 15:16:55.050090] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid81670 ] 00:11:05.107 [2024-11-19 15:16:55.205896] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:05.107 [2024-11-19 15:16:55.245160] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:05.107 [2024-11-19 15:16:55.321504] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:05.107 [2024-11-19 15:16:55.321653] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:05.678 15:16:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:05.678 15:16:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:11:05.678 15:16:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:11:05.678 15:16:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:11:05.678 15:16:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:11:05.678 15:16:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:11:05.678 15:16:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:11:05.678 15:16:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:11:05.678 15:16:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:11:05.678 15:16:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:11:05.678 15:16:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:11:05.678 15:16:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:05.678 15:16:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.678 malloc1 00:11:05.678 15:16:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:05.678 15:16:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:11:05.678 15:16:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:05.678 15:16:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.678 [2024-11-19 15:16:55.959869] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:11:05.678 [2024-11-19 15:16:55.959939] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:05.678 [2024-11-19 15:16:55.959961] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:11:05.678 [2024-11-19 15:16:55.959989] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:05.678 [2024-11-19 15:16:55.962398] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:05.678 [2024-11-19 15:16:55.962437] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:11:05.678 pt1 00:11:05.678 15:16:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:05.678 15:16:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:11:05.678 15:16:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:11:05.678 15:16:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:11:05.678 15:16:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:11:05.678 15:16:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:11:05.678 15:16:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:11:05.678 15:16:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:11:05.678 15:16:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:11:05.678 15:16:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:11:05.678 15:16:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:05.678 15:16:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.678 malloc2 00:11:05.678 15:16:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:05.678 15:16:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:11:05.678 15:16:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:05.678 15:16:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.678 [2024-11-19 15:16:55.994466] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:11:05.678 [2024-11-19 15:16:55.994594] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:05.678 [2024-11-19 15:16:55.994628] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:11:05.678 [2024-11-19 15:16:55.994664] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:05.678 [2024-11-19 15:16:55.997100] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:05.678 [2024-11-19 15:16:55.997171] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:11:05.678 pt2 00:11:05.678 15:16:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:05.678 15:16:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:11:05.678 15:16:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:11:05.678 15:16:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:11:05.678 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:11:05.678 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:11:05.678 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:11:05.678 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:11:05.678 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:11:05.678 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:11:05.678 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:05.678 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.939 malloc3 00:11:05.939 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:05.939 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:11:05.939 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:05.939 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.939 [2024-11-19 15:16:56.033189] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:11:05.939 [2024-11-19 15:16:56.033308] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:05.939 [2024-11-19 15:16:56.033348] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:11:05.939 [2024-11-19 15:16:56.033379] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:05.939 [2024-11-19 15:16:56.035772] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:05.939 [2024-11-19 15:16:56.035848] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:11:05.939 pt3 00:11:05.939 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:05.939 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:11:05.939 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:11:05.939 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:11:05.939 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:11:05.939 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:11:05.939 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:11:05.939 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:11:05.939 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:11:05.939 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:11:05.939 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:05.939 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.939 malloc4 00:11:05.939 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:05.939 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:11:05.939 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:05.939 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.939 [2024-11-19 15:16:56.080259] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:11:05.939 [2024-11-19 15:16:56.080373] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:05.939 [2024-11-19 15:16:56.080392] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:11:05.939 [2024-11-19 15:16:56.080407] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:05.939 [2024-11-19 15:16:56.082780] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:05.939 [2024-11-19 15:16:56.082817] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:11:05.939 pt4 00:11:05.940 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:05.940 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:11:05.940 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:11:05.940 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:11:05.940 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:05.940 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.940 [2024-11-19 15:16:56.092260] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:11:05.940 [2024-11-19 15:16:56.094380] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:11:05.940 [2024-11-19 15:16:56.094449] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:11:05.940 [2024-11-19 15:16:56.094494] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:11:05.940 [2024-11-19 15:16:56.094648] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:11:05.940 [2024-11-19 15:16:56.094661] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:11:05.940 [2024-11-19 15:16:56.094902] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:11:05.940 [2024-11-19 15:16:56.095066] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:11:05.940 [2024-11-19 15:16:56.095077] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:11:05.940 [2024-11-19 15:16:56.095207] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:05.940 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:05.940 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:11:05.940 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:05.940 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:05.940 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:11:05.940 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:05.940 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:05.940 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:05.940 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:05.940 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:05.940 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:05.940 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:05.940 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:05.940 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:05.940 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.940 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:05.940 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:05.940 "name": "raid_bdev1", 00:11:05.940 "uuid": "88a85a64-870f-4f88-bc70-a4a4db4a2cc5", 00:11:05.940 "strip_size_kb": 64, 00:11:05.940 "state": "online", 00:11:05.940 "raid_level": "raid0", 00:11:05.940 "superblock": true, 00:11:05.940 "num_base_bdevs": 4, 00:11:05.940 "num_base_bdevs_discovered": 4, 00:11:05.940 "num_base_bdevs_operational": 4, 00:11:05.940 "base_bdevs_list": [ 00:11:05.940 { 00:11:05.940 "name": "pt1", 00:11:05.940 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:05.940 "is_configured": true, 00:11:05.940 "data_offset": 2048, 00:11:05.940 "data_size": 63488 00:11:05.940 }, 00:11:05.940 { 00:11:05.940 "name": "pt2", 00:11:05.940 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:05.940 "is_configured": true, 00:11:05.940 "data_offset": 2048, 00:11:05.940 "data_size": 63488 00:11:05.940 }, 00:11:05.940 { 00:11:05.940 "name": "pt3", 00:11:05.940 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:05.940 "is_configured": true, 00:11:05.940 "data_offset": 2048, 00:11:05.940 "data_size": 63488 00:11:05.940 }, 00:11:05.940 { 00:11:05.940 "name": "pt4", 00:11:05.940 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:05.940 "is_configured": true, 00:11:05.940 "data_offset": 2048, 00:11:05.940 "data_size": 63488 00:11:05.940 } 00:11:05.940 ] 00:11:05.940 }' 00:11:05.940 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:05.940 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.210 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:11:06.210 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:11:06.210 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:11:06.210 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:11:06.210 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:11:06.210 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:11:06.210 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:11:06.210 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:06.210 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.210 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.210 [2024-11-19 15:16:56.527971] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:06.488 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.488 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:11:06.488 "name": "raid_bdev1", 00:11:06.488 "aliases": [ 00:11:06.488 "88a85a64-870f-4f88-bc70-a4a4db4a2cc5" 00:11:06.488 ], 00:11:06.488 "product_name": "Raid Volume", 00:11:06.488 "block_size": 512, 00:11:06.488 "num_blocks": 253952, 00:11:06.488 "uuid": "88a85a64-870f-4f88-bc70-a4a4db4a2cc5", 00:11:06.488 "assigned_rate_limits": { 00:11:06.488 "rw_ios_per_sec": 0, 00:11:06.488 "rw_mbytes_per_sec": 0, 00:11:06.488 "r_mbytes_per_sec": 0, 00:11:06.488 "w_mbytes_per_sec": 0 00:11:06.488 }, 00:11:06.488 "claimed": false, 00:11:06.488 "zoned": false, 00:11:06.488 "supported_io_types": { 00:11:06.488 "read": true, 00:11:06.488 "write": true, 00:11:06.488 "unmap": true, 00:11:06.488 "flush": true, 00:11:06.488 "reset": true, 00:11:06.488 "nvme_admin": false, 00:11:06.488 "nvme_io": false, 00:11:06.488 "nvme_io_md": false, 00:11:06.488 "write_zeroes": true, 00:11:06.488 "zcopy": false, 00:11:06.488 "get_zone_info": false, 00:11:06.488 "zone_management": false, 00:11:06.488 "zone_append": false, 00:11:06.488 "compare": false, 00:11:06.488 "compare_and_write": false, 00:11:06.488 "abort": false, 00:11:06.488 "seek_hole": false, 00:11:06.488 "seek_data": false, 00:11:06.488 "copy": false, 00:11:06.488 "nvme_iov_md": false 00:11:06.488 }, 00:11:06.489 "memory_domains": [ 00:11:06.489 { 00:11:06.489 "dma_device_id": "system", 00:11:06.489 "dma_device_type": 1 00:11:06.489 }, 00:11:06.489 { 00:11:06.489 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:06.489 "dma_device_type": 2 00:11:06.489 }, 00:11:06.489 { 00:11:06.489 "dma_device_id": "system", 00:11:06.489 "dma_device_type": 1 00:11:06.489 }, 00:11:06.489 { 00:11:06.489 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:06.489 "dma_device_type": 2 00:11:06.489 }, 00:11:06.489 { 00:11:06.489 "dma_device_id": "system", 00:11:06.489 "dma_device_type": 1 00:11:06.489 }, 00:11:06.489 { 00:11:06.489 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:06.489 "dma_device_type": 2 00:11:06.489 }, 00:11:06.489 { 00:11:06.489 "dma_device_id": "system", 00:11:06.489 "dma_device_type": 1 00:11:06.489 }, 00:11:06.489 { 00:11:06.489 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:06.489 "dma_device_type": 2 00:11:06.489 } 00:11:06.489 ], 00:11:06.489 "driver_specific": { 00:11:06.489 "raid": { 00:11:06.489 "uuid": "88a85a64-870f-4f88-bc70-a4a4db4a2cc5", 00:11:06.489 "strip_size_kb": 64, 00:11:06.489 "state": "online", 00:11:06.489 "raid_level": "raid0", 00:11:06.489 "superblock": true, 00:11:06.489 "num_base_bdevs": 4, 00:11:06.489 "num_base_bdevs_discovered": 4, 00:11:06.489 "num_base_bdevs_operational": 4, 00:11:06.489 "base_bdevs_list": [ 00:11:06.489 { 00:11:06.489 "name": "pt1", 00:11:06.489 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:06.489 "is_configured": true, 00:11:06.489 "data_offset": 2048, 00:11:06.489 "data_size": 63488 00:11:06.489 }, 00:11:06.489 { 00:11:06.489 "name": "pt2", 00:11:06.489 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:06.489 "is_configured": true, 00:11:06.489 "data_offset": 2048, 00:11:06.489 "data_size": 63488 00:11:06.489 }, 00:11:06.489 { 00:11:06.489 "name": "pt3", 00:11:06.489 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:06.489 "is_configured": true, 00:11:06.489 "data_offset": 2048, 00:11:06.489 "data_size": 63488 00:11:06.489 }, 00:11:06.489 { 00:11:06.489 "name": "pt4", 00:11:06.489 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:06.489 "is_configured": true, 00:11:06.489 "data_offset": 2048, 00:11:06.489 "data_size": 63488 00:11:06.489 } 00:11:06.489 ] 00:11:06.489 } 00:11:06.489 } 00:11:06.489 }' 00:11:06.489 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:11:06.489 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:11:06.489 pt2 00:11:06.489 pt3 00:11:06.489 pt4' 00:11:06.489 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:06.489 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:11:06.489 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:06.489 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:11:06.489 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.489 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.489 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:06.489 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.489 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:06.489 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:06.489 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:06.489 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:11:06.489 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.489 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.489 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:06.489 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.489 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:06.489 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:06.489 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:06.489 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:11:06.489 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.489 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.489 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:06.489 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.489 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:06.489 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:06.489 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:06.489 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:06.489 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:11:06.489 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.489 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.489 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.489 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:06.489 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:06.489 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:11:06.489 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:06.489 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.489 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.489 [2024-11-19 15:16:56.815359] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:06.750 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.750 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=88a85a64-870f-4f88-bc70-a4a4db4a2cc5 00:11:06.750 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 88a85a64-870f-4f88-bc70-a4a4db4a2cc5 ']' 00:11:06.750 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:06.750 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.750 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.750 [2024-11-19 15:16:56.851039] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:06.750 [2024-11-19 15:16:56.851076] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:06.750 [2024-11-19 15:16:56.851162] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:06.750 [2024-11-19 15:16:56.851243] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:06.750 [2024-11-19 15:16:56.851253] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:11:06.750 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.750 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:06.750 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.750 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.750 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:11:06.750 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.750 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:11:06.750 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:11:06.750 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:11:06.750 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:11:06.750 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.750 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.750 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.750 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:11:06.750 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:11:06.750 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.750 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.750 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.750 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:11:06.750 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:11:06.751 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.751 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.751 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.751 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:11:06.751 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:11:06.751 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.751 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.751 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.751 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:11:06.751 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.751 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.751 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:11:06.751 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.751 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:11:06.751 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:11:06.751 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:11:06.751 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:11:06.751 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:11:06.751 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:11:06.751 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:11:06.751 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:11:06.751 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:11:06.751 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.751 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.751 [2024-11-19 15:16:56.978863] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:11:06.751 [2024-11-19 15:16:56.981180] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:11:06.751 [2024-11-19 15:16:56.981232] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:11:06.751 [2024-11-19 15:16:56.981263] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:11:06.751 [2024-11-19 15:16:56.981314] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:11:06.751 [2024-11-19 15:16:56.981373] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:11:06.751 [2024-11-19 15:16:56.981395] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:11:06.751 [2024-11-19 15:16:56.981411] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:11:06.751 [2024-11-19 15:16:56.981427] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:06.751 [2024-11-19 15:16:56.981437] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:11:06.751 request: 00:11:06.751 { 00:11:06.751 "name": "raid_bdev1", 00:11:06.751 "raid_level": "raid0", 00:11:06.751 "base_bdevs": [ 00:11:06.751 "malloc1", 00:11:06.751 "malloc2", 00:11:06.751 "malloc3", 00:11:06.751 "malloc4" 00:11:06.751 ], 00:11:06.751 "strip_size_kb": 64, 00:11:06.751 "superblock": false, 00:11:06.751 "method": "bdev_raid_create", 00:11:06.751 "req_id": 1 00:11:06.751 } 00:11:06.751 Got JSON-RPC error response 00:11:06.751 response: 00:11:06.751 { 00:11:06.751 "code": -17, 00:11:06.751 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:11:06.751 } 00:11:06.751 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:11:06.751 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:11:06.751 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:11:06.751 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:11:06.751 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:11:06.751 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:06.751 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.751 15:16:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.751 15:16:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:11:06.751 15:16:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.751 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:11:06.751 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:11:06.751 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:11:06.751 15:16:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.751 15:16:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.751 [2024-11-19 15:16:57.030712] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:11:06.751 [2024-11-19 15:16:57.030841] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:06.751 [2024-11-19 15:16:57.030882] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:11:06.751 [2024-11-19 15:16:57.030912] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:06.751 [2024-11-19 15:16:57.033450] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:06.751 [2024-11-19 15:16:57.033522] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:11:06.751 [2024-11-19 15:16:57.033621] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:11:06.751 [2024-11-19 15:16:57.033684] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:11:06.751 pt1 00:11:06.751 15:16:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.751 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 4 00:11:06.751 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:06.751 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:06.751 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:11:06.751 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:06.751 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:06.751 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:06.751 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:06.751 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:06.751 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:06.751 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:06.751 15:16:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.751 15:16:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.751 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:06.751 15:16:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:07.012 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:07.012 "name": "raid_bdev1", 00:11:07.012 "uuid": "88a85a64-870f-4f88-bc70-a4a4db4a2cc5", 00:11:07.012 "strip_size_kb": 64, 00:11:07.012 "state": "configuring", 00:11:07.012 "raid_level": "raid0", 00:11:07.012 "superblock": true, 00:11:07.012 "num_base_bdevs": 4, 00:11:07.012 "num_base_bdevs_discovered": 1, 00:11:07.012 "num_base_bdevs_operational": 4, 00:11:07.012 "base_bdevs_list": [ 00:11:07.012 { 00:11:07.012 "name": "pt1", 00:11:07.012 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:07.012 "is_configured": true, 00:11:07.012 "data_offset": 2048, 00:11:07.012 "data_size": 63488 00:11:07.012 }, 00:11:07.012 { 00:11:07.012 "name": null, 00:11:07.012 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:07.012 "is_configured": false, 00:11:07.012 "data_offset": 2048, 00:11:07.012 "data_size": 63488 00:11:07.012 }, 00:11:07.012 { 00:11:07.012 "name": null, 00:11:07.012 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:07.012 "is_configured": false, 00:11:07.012 "data_offset": 2048, 00:11:07.012 "data_size": 63488 00:11:07.012 }, 00:11:07.012 { 00:11:07.012 "name": null, 00:11:07.012 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:07.012 "is_configured": false, 00:11:07.012 "data_offset": 2048, 00:11:07.012 "data_size": 63488 00:11:07.012 } 00:11:07.012 ] 00:11:07.012 }' 00:11:07.012 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:07.012 15:16:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:07.273 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:11:07.273 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:11:07.273 15:16:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:07.273 15:16:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:07.273 [2024-11-19 15:16:57.446071] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:11:07.273 [2024-11-19 15:16:57.446236] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:07.273 [2024-11-19 15:16:57.446278] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:11:07.273 [2024-11-19 15:16:57.446319] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:07.273 [2024-11-19 15:16:57.446828] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:07.273 [2024-11-19 15:16:57.446887] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:11:07.273 [2024-11-19 15:16:57.447020] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:11:07.273 [2024-11-19 15:16:57.447073] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:11:07.273 pt2 00:11:07.273 15:16:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:07.273 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:11:07.273 15:16:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:07.273 15:16:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:07.273 [2024-11-19 15:16:57.454051] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:11:07.273 15:16:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:07.273 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 4 00:11:07.273 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:07.273 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:07.273 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:11:07.273 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:07.273 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:07.273 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:07.273 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:07.273 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:07.273 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:07.273 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:07.273 15:16:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:07.273 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:07.273 15:16:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:07.273 15:16:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:07.273 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:07.273 "name": "raid_bdev1", 00:11:07.273 "uuid": "88a85a64-870f-4f88-bc70-a4a4db4a2cc5", 00:11:07.273 "strip_size_kb": 64, 00:11:07.273 "state": "configuring", 00:11:07.273 "raid_level": "raid0", 00:11:07.273 "superblock": true, 00:11:07.273 "num_base_bdevs": 4, 00:11:07.273 "num_base_bdevs_discovered": 1, 00:11:07.273 "num_base_bdevs_operational": 4, 00:11:07.273 "base_bdevs_list": [ 00:11:07.273 { 00:11:07.273 "name": "pt1", 00:11:07.273 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:07.273 "is_configured": true, 00:11:07.273 "data_offset": 2048, 00:11:07.273 "data_size": 63488 00:11:07.273 }, 00:11:07.273 { 00:11:07.273 "name": null, 00:11:07.273 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:07.273 "is_configured": false, 00:11:07.273 "data_offset": 0, 00:11:07.273 "data_size": 63488 00:11:07.273 }, 00:11:07.273 { 00:11:07.273 "name": null, 00:11:07.273 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:07.273 "is_configured": false, 00:11:07.273 "data_offset": 2048, 00:11:07.273 "data_size": 63488 00:11:07.273 }, 00:11:07.273 { 00:11:07.273 "name": null, 00:11:07.273 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:07.273 "is_configured": false, 00:11:07.273 "data_offset": 2048, 00:11:07.273 "data_size": 63488 00:11:07.273 } 00:11:07.273 ] 00:11:07.273 }' 00:11:07.284 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:07.284 15:16:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:07.545 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:11:07.545 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:11:07.545 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:11:07.545 15:16:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:07.545 15:16:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:07.545 [2024-11-19 15:16:57.881370] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:11:07.545 [2024-11-19 15:16:57.881575] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:07.545 [2024-11-19 15:16:57.881617] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:11:07.545 [2024-11-19 15:16:57.881654] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:07.545 [2024-11-19 15:16:57.882185] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:07.545 [2024-11-19 15:16:57.882248] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:11:07.545 [2024-11-19 15:16:57.882378] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:11:07.545 [2024-11-19 15:16:57.882434] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:11:07.807 pt2 00:11:07.807 15:16:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:07.807 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:11:07.807 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:11:07.807 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:11:07.807 15:16:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:07.807 15:16:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:07.807 [2024-11-19 15:16:57.893277] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:11:07.807 [2024-11-19 15:16:57.893418] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:07.807 [2024-11-19 15:16:57.893459] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:11:07.807 [2024-11-19 15:16:57.893493] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:07.807 [2024-11-19 15:16:57.894042] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:07.807 [2024-11-19 15:16:57.894109] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:11:07.807 [2024-11-19 15:16:57.894229] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:11:07.807 [2024-11-19 15:16:57.894284] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:11:07.807 pt3 00:11:07.807 15:16:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:07.807 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:11:07.807 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:11:07.807 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:11:07.807 15:16:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:07.807 15:16:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:07.807 [2024-11-19 15:16:57.905213] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:11:07.807 [2024-11-19 15:16:57.905272] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:07.807 [2024-11-19 15:16:57.905288] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:11:07.807 [2024-11-19 15:16:57.905298] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:07.807 [2024-11-19 15:16:57.905650] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:07.807 [2024-11-19 15:16:57.905669] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:11:07.807 [2024-11-19 15:16:57.905731] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:11:07.807 [2024-11-19 15:16:57.905752] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:11:07.807 [2024-11-19 15:16:57.905854] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:11:07.807 [2024-11-19 15:16:57.905866] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:11:07.807 [2024-11-19 15:16:57.906138] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:11:07.807 [2024-11-19 15:16:57.906272] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:11:07.807 [2024-11-19 15:16:57.906281] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:11:07.807 [2024-11-19 15:16:57.906385] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:07.807 pt4 00:11:07.807 15:16:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:07.807 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:11:07.807 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:11:07.807 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:11:07.807 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:07.807 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:07.807 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:11:07.807 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:07.807 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:07.807 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:07.807 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:07.807 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:07.807 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:07.807 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:07.807 15:16:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:07.807 15:16:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:07.807 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:07.807 15:16:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:07.807 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:07.807 "name": "raid_bdev1", 00:11:07.807 "uuid": "88a85a64-870f-4f88-bc70-a4a4db4a2cc5", 00:11:07.807 "strip_size_kb": 64, 00:11:07.807 "state": "online", 00:11:07.807 "raid_level": "raid0", 00:11:07.807 "superblock": true, 00:11:07.807 "num_base_bdevs": 4, 00:11:07.807 "num_base_bdevs_discovered": 4, 00:11:07.807 "num_base_bdevs_operational": 4, 00:11:07.807 "base_bdevs_list": [ 00:11:07.807 { 00:11:07.807 "name": "pt1", 00:11:07.807 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:07.807 "is_configured": true, 00:11:07.807 "data_offset": 2048, 00:11:07.807 "data_size": 63488 00:11:07.807 }, 00:11:07.807 { 00:11:07.807 "name": "pt2", 00:11:07.807 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:07.808 "is_configured": true, 00:11:07.808 "data_offset": 2048, 00:11:07.808 "data_size": 63488 00:11:07.808 }, 00:11:07.808 { 00:11:07.808 "name": "pt3", 00:11:07.808 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:07.808 "is_configured": true, 00:11:07.808 "data_offset": 2048, 00:11:07.808 "data_size": 63488 00:11:07.808 }, 00:11:07.808 { 00:11:07.808 "name": "pt4", 00:11:07.808 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:07.808 "is_configured": true, 00:11:07.808 "data_offset": 2048, 00:11:07.808 "data_size": 63488 00:11:07.808 } 00:11:07.808 ] 00:11:07.808 }' 00:11:07.808 15:16:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:07.808 15:16:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.069 15:16:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:11:08.069 15:16:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:11:08.069 15:16:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:11:08.069 15:16:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:11:08.069 15:16:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:11:08.069 15:16:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:11:08.069 15:16:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:08.069 15:16:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:08.069 15:16:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.069 15:16:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:11:08.069 [2024-11-19 15:16:58.340869] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:08.069 15:16:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:08.069 15:16:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:11:08.069 "name": "raid_bdev1", 00:11:08.069 "aliases": [ 00:11:08.069 "88a85a64-870f-4f88-bc70-a4a4db4a2cc5" 00:11:08.069 ], 00:11:08.069 "product_name": "Raid Volume", 00:11:08.069 "block_size": 512, 00:11:08.069 "num_blocks": 253952, 00:11:08.069 "uuid": "88a85a64-870f-4f88-bc70-a4a4db4a2cc5", 00:11:08.069 "assigned_rate_limits": { 00:11:08.069 "rw_ios_per_sec": 0, 00:11:08.069 "rw_mbytes_per_sec": 0, 00:11:08.069 "r_mbytes_per_sec": 0, 00:11:08.069 "w_mbytes_per_sec": 0 00:11:08.069 }, 00:11:08.069 "claimed": false, 00:11:08.069 "zoned": false, 00:11:08.069 "supported_io_types": { 00:11:08.069 "read": true, 00:11:08.069 "write": true, 00:11:08.069 "unmap": true, 00:11:08.069 "flush": true, 00:11:08.069 "reset": true, 00:11:08.069 "nvme_admin": false, 00:11:08.069 "nvme_io": false, 00:11:08.069 "nvme_io_md": false, 00:11:08.069 "write_zeroes": true, 00:11:08.069 "zcopy": false, 00:11:08.069 "get_zone_info": false, 00:11:08.069 "zone_management": false, 00:11:08.069 "zone_append": false, 00:11:08.069 "compare": false, 00:11:08.069 "compare_and_write": false, 00:11:08.069 "abort": false, 00:11:08.069 "seek_hole": false, 00:11:08.069 "seek_data": false, 00:11:08.069 "copy": false, 00:11:08.069 "nvme_iov_md": false 00:11:08.069 }, 00:11:08.069 "memory_domains": [ 00:11:08.069 { 00:11:08.069 "dma_device_id": "system", 00:11:08.069 "dma_device_type": 1 00:11:08.069 }, 00:11:08.069 { 00:11:08.069 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:08.069 "dma_device_type": 2 00:11:08.069 }, 00:11:08.069 { 00:11:08.069 "dma_device_id": "system", 00:11:08.069 "dma_device_type": 1 00:11:08.069 }, 00:11:08.069 { 00:11:08.069 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:08.069 "dma_device_type": 2 00:11:08.069 }, 00:11:08.069 { 00:11:08.069 "dma_device_id": "system", 00:11:08.069 "dma_device_type": 1 00:11:08.069 }, 00:11:08.069 { 00:11:08.069 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:08.069 "dma_device_type": 2 00:11:08.069 }, 00:11:08.069 { 00:11:08.069 "dma_device_id": "system", 00:11:08.069 "dma_device_type": 1 00:11:08.069 }, 00:11:08.069 { 00:11:08.069 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:08.069 "dma_device_type": 2 00:11:08.069 } 00:11:08.069 ], 00:11:08.069 "driver_specific": { 00:11:08.069 "raid": { 00:11:08.069 "uuid": "88a85a64-870f-4f88-bc70-a4a4db4a2cc5", 00:11:08.069 "strip_size_kb": 64, 00:11:08.069 "state": "online", 00:11:08.069 "raid_level": "raid0", 00:11:08.069 "superblock": true, 00:11:08.069 "num_base_bdevs": 4, 00:11:08.069 "num_base_bdevs_discovered": 4, 00:11:08.069 "num_base_bdevs_operational": 4, 00:11:08.069 "base_bdevs_list": [ 00:11:08.069 { 00:11:08.069 "name": "pt1", 00:11:08.069 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:08.069 "is_configured": true, 00:11:08.069 "data_offset": 2048, 00:11:08.069 "data_size": 63488 00:11:08.069 }, 00:11:08.069 { 00:11:08.069 "name": "pt2", 00:11:08.069 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:08.069 "is_configured": true, 00:11:08.069 "data_offset": 2048, 00:11:08.069 "data_size": 63488 00:11:08.069 }, 00:11:08.069 { 00:11:08.069 "name": "pt3", 00:11:08.069 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:08.069 "is_configured": true, 00:11:08.069 "data_offset": 2048, 00:11:08.069 "data_size": 63488 00:11:08.069 }, 00:11:08.069 { 00:11:08.069 "name": "pt4", 00:11:08.069 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:08.069 "is_configured": true, 00:11:08.069 "data_offset": 2048, 00:11:08.069 "data_size": 63488 00:11:08.069 } 00:11:08.069 ] 00:11:08.069 } 00:11:08.069 } 00:11:08.069 }' 00:11:08.069 15:16:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:11:08.329 15:16:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:11:08.329 pt2 00:11:08.329 pt3 00:11:08.329 pt4' 00:11:08.329 15:16:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:08.329 15:16:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:11:08.329 15:16:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:08.329 15:16:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:11:08.329 15:16:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:08.329 15:16:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.329 15:16:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:08.329 15:16:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:08.329 15:16:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:08.329 15:16:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:08.329 15:16:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:08.329 15:16:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:11:08.329 15:16:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:08.329 15:16:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.329 15:16:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:08.329 15:16:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:08.329 15:16:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:08.329 15:16:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:08.329 15:16:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:08.329 15:16:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:11:08.330 15:16:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:08.330 15:16:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:08.330 15:16:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.330 15:16:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:08.330 15:16:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:08.330 15:16:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:08.330 15:16:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:08.330 15:16:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:11:08.330 15:16:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:08.330 15:16:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:08.330 15:16:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.330 15:16:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:08.590 15:16:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:08.590 15:16:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:08.590 15:16:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:08.590 15:16:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:08.590 15:16:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.590 15:16:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:11:08.590 [2024-11-19 15:16:58.680281] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:08.590 15:16:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:08.590 15:16:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 88a85a64-870f-4f88-bc70-a4a4db4a2cc5 '!=' 88a85a64-870f-4f88-bc70-a4a4db4a2cc5 ']' 00:11:08.590 15:16:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid0 00:11:08.590 15:16:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:11:08.590 15:16:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:11:08.590 15:16:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 81670 00:11:08.590 15:16:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 81670 ']' 00:11:08.590 15:16:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 81670 00:11:08.590 15:16:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:11:08.590 15:16:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:11:08.590 15:16:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 81670 00:11:08.590 killing process with pid 81670 00:11:08.590 15:16:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:11:08.590 15:16:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:11:08.590 15:16:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 81670' 00:11:08.590 15:16:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 81670 00:11:08.590 [2024-11-19 15:16:58.766885] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:08.590 [2024-11-19 15:16:58.767033] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:08.590 15:16:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 81670 00:11:08.590 [2024-11-19 15:16:58.767126] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:08.590 [2024-11-19 15:16:58.767141] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:11:08.590 [2024-11-19 15:16:58.845947] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:08.850 15:16:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:11:08.850 00:11:08.850 real 0m4.197s 00:11:08.850 user 0m6.377s 00:11:08.850 sys 0m0.965s 00:11:08.850 15:16:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:08.850 ************************************ 00:11:08.850 END TEST raid_superblock_test 00:11:08.850 ************************************ 00:11:08.850 15:16:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:09.109 15:16:59 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid0 4 read 00:11:09.109 15:16:59 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:11:09.109 15:16:59 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:09.109 15:16:59 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:09.109 ************************************ 00:11:09.109 START TEST raid_read_error_test 00:11:09.109 ************************************ 00:11:09.109 15:16:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid0 4 read 00:11:09.109 15:16:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:11:09.109 15:16:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:11:09.109 15:16:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:11:09.109 15:16:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:11:09.109 15:16:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:09.109 15:16:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:11:09.109 15:16:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:09.109 15:16:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:09.109 15:16:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:11:09.109 15:16:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:09.109 15:16:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:09.109 15:16:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:11:09.109 15:16:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:09.109 15:16:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:09.109 15:16:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:11:09.109 15:16:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:09.109 15:16:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:09.109 15:16:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:11:09.109 15:16:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:11:09.109 15:16:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:11:09.109 15:16:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:11:09.109 15:16:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:11:09.109 15:16:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:11:09.109 15:16:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:11:09.109 15:16:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:11:09.109 15:16:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:11:09.109 15:16:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:11:09.109 15:16:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:11:09.110 15:16:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.0aldQlrCxa 00:11:09.110 15:16:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=81918 00:11:09.110 15:16:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:11:09.110 15:16:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 81918 00:11:09.110 15:16:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 81918 ']' 00:11:09.110 15:16:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:09.110 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:09.110 15:16:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:09.110 15:16:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:09.110 15:16:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:09.110 15:16:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:09.110 [2024-11-19 15:16:59.356439] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:11:09.110 [2024-11-19 15:16:59.356646] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid81918 ] 00:11:09.369 [2024-11-19 15:16:59.510727] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:09.369 [2024-11-19 15:16:59.550622] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:09.369 [2024-11-19 15:16:59.626672] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:09.369 [2024-11-19 15:16:59.626720] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:09.939 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:09.939 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:11:09.939 15:17:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:09.939 15:17:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:11:09.939 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:09.939 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:09.939 BaseBdev1_malloc 00:11:09.939 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:09.939 15:17:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:11:09.939 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:09.939 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:09.939 true 00:11:09.939 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:09.939 15:17:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:11:09.939 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:09.939 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:09.939 [2024-11-19 15:17:00.233090] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:11:09.939 [2024-11-19 15:17:00.233161] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:09.939 [2024-11-19 15:17:00.233184] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:11:09.939 [2024-11-19 15:17:00.233193] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:09.939 [2024-11-19 15:17:00.235690] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:09.939 BaseBdev1 00:11:09.939 [2024-11-19 15:17:00.235824] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:09.939 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:09.939 15:17:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:09.939 15:17:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:11:09.939 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:09.939 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:09.939 BaseBdev2_malloc 00:11:09.939 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:09.939 15:17:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:11:09.939 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:09.939 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:09.939 true 00:11:09.939 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:09.939 15:17:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:11:09.939 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:09.939 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:10.200 [2024-11-19 15:17:00.279844] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:11:10.200 [2024-11-19 15:17:00.279903] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:10.200 [2024-11-19 15:17:00.279924] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:11:10.200 [2024-11-19 15:17:00.279943] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:10.200 [2024-11-19 15:17:00.282460] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:10.200 [2024-11-19 15:17:00.282500] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:11:10.200 BaseBdev2 00:11:10.200 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:10.200 15:17:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:10.200 15:17:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:11:10.200 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:10.200 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:10.200 BaseBdev3_malloc 00:11:10.200 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:10.200 15:17:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:11:10.200 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:10.200 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:10.200 true 00:11:10.200 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:10.200 15:17:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:11:10.200 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:10.200 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:10.200 [2024-11-19 15:17:00.326569] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:11:10.200 [2024-11-19 15:17:00.326627] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:10.200 [2024-11-19 15:17:00.326647] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:11:10.200 [2024-11-19 15:17:00.326657] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:10.200 [2024-11-19 15:17:00.329131] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:10.200 [2024-11-19 15:17:00.329167] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:11:10.200 BaseBdev3 00:11:10.200 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:10.200 15:17:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:10.200 15:17:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:11:10.200 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:10.200 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:10.200 BaseBdev4_malloc 00:11:10.200 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:10.200 15:17:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:11:10.200 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:10.200 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:10.200 true 00:11:10.200 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:10.200 15:17:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:11:10.200 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:10.200 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:10.200 [2024-11-19 15:17:00.381830] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:11:10.200 [2024-11-19 15:17:00.381982] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:10.200 [2024-11-19 15:17:00.382016] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:11:10.200 [2024-11-19 15:17:00.382026] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:10.200 [2024-11-19 15:17:00.384511] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:10.200 [2024-11-19 15:17:00.384552] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:11:10.200 BaseBdev4 00:11:10.200 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:10.200 15:17:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:11:10.200 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:10.200 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:10.200 [2024-11-19 15:17:00.393867] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:10.200 [2024-11-19 15:17:00.396018] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:10.200 [2024-11-19 15:17:00.396163] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:10.200 [2024-11-19 15:17:00.396230] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:10.200 [2024-11-19 15:17:00.396445] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002000 00:11:10.201 [2024-11-19 15:17:00.396465] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:11:10.201 [2024-11-19 15:17:00.396722] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002ef0 00:11:10.201 [2024-11-19 15:17:00.396868] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002000 00:11:10.201 [2024-11-19 15:17:00.396881] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002000 00:11:10.201 [2024-11-19 15:17:00.397040] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:10.201 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:10.201 15:17:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:11:10.201 15:17:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:10.201 15:17:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:10.201 15:17:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:11:10.201 15:17:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:10.201 15:17:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:10.201 15:17:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:10.201 15:17:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:10.201 15:17:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:10.201 15:17:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:10.201 15:17:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:10.201 15:17:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:10.201 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:10.201 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:10.201 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:10.201 15:17:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:10.201 "name": "raid_bdev1", 00:11:10.201 "uuid": "cf62be57-f425-459f-8632-ee71357cb404", 00:11:10.201 "strip_size_kb": 64, 00:11:10.201 "state": "online", 00:11:10.201 "raid_level": "raid0", 00:11:10.201 "superblock": true, 00:11:10.201 "num_base_bdevs": 4, 00:11:10.201 "num_base_bdevs_discovered": 4, 00:11:10.201 "num_base_bdevs_operational": 4, 00:11:10.201 "base_bdevs_list": [ 00:11:10.201 { 00:11:10.201 "name": "BaseBdev1", 00:11:10.201 "uuid": "ccbdef4f-9b0c-5ab7-8e0d-e3ea3e804443", 00:11:10.201 "is_configured": true, 00:11:10.201 "data_offset": 2048, 00:11:10.201 "data_size": 63488 00:11:10.201 }, 00:11:10.201 { 00:11:10.201 "name": "BaseBdev2", 00:11:10.201 "uuid": "846bc7e5-71ee-5bec-a704-86a3f1e4d873", 00:11:10.201 "is_configured": true, 00:11:10.201 "data_offset": 2048, 00:11:10.201 "data_size": 63488 00:11:10.201 }, 00:11:10.201 { 00:11:10.201 "name": "BaseBdev3", 00:11:10.201 "uuid": "a01325a4-5c8f-5187-b703-ced4db8f999d", 00:11:10.201 "is_configured": true, 00:11:10.201 "data_offset": 2048, 00:11:10.201 "data_size": 63488 00:11:10.201 }, 00:11:10.201 { 00:11:10.201 "name": "BaseBdev4", 00:11:10.201 "uuid": "5983d9fb-0f9a-56c5-b8b2-501f3593111c", 00:11:10.201 "is_configured": true, 00:11:10.201 "data_offset": 2048, 00:11:10.201 "data_size": 63488 00:11:10.201 } 00:11:10.201 ] 00:11:10.201 }' 00:11:10.201 15:17:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:10.201 15:17:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:10.769 15:17:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:11:10.769 15:17:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:11:10.769 [2024-11-19 15:17:00.901612] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000003090 00:11:11.706 15:17:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:11:11.706 15:17:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:11.706 15:17:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:11.706 15:17:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:11.706 15:17:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:11:11.706 15:17:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:11:11.706 15:17:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:11:11.706 15:17:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:11:11.706 15:17:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:11.706 15:17:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:11.706 15:17:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:11:11.706 15:17:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:11.706 15:17:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:11.706 15:17:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:11.706 15:17:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:11.706 15:17:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:11.706 15:17:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:11.706 15:17:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:11.706 15:17:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:11.706 15:17:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:11.706 15:17:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:11.706 15:17:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:11.706 15:17:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:11.706 "name": "raid_bdev1", 00:11:11.706 "uuid": "cf62be57-f425-459f-8632-ee71357cb404", 00:11:11.706 "strip_size_kb": 64, 00:11:11.706 "state": "online", 00:11:11.706 "raid_level": "raid0", 00:11:11.706 "superblock": true, 00:11:11.706 "num_base_bdevs": 4, 00:11:11.706 "num_base_bdevs_discovered": 4, 00:11:11.706 "num_base_bdevs_operational": 4, 00:11:11.706 "base_bdevs_list": [ 00:11:11.706 { 00:11:11.706 "name": "BaseBdev1", 00:11:11.706 "uuid": "ccbdef4f-9b0c-5ab7-8e0d-e3ea3e804443", 00:11:11.706 "is_configured": true, 00:11:11.706 "data_offset": 2048, 00:11:11.706 "data_size": 63488 00:11:11.706 }, 00:11:11.706 { 00:11:11.706 "name": "BaseBdev2", 00:11:11.706 "uuid": "846bc7e5-71ee-5bec-a704-86a3f1e4d873", 00:11:11.706 "is_configured": true, 00:11:11.706 "data_offset": 2048, 00:11:11.706 "data_size": 63488 00:11:11.706 }, 00:11:11.706 { 00:11:11.706 "name": "BaseBdev3", 00:11:11.706 "uuid": "a01325a4-5c8f-5187-b703-ced4db8f999d", 00:11:11.706 "is_configured": true, 00:11:11.706 "data_offset": 2048, 00:11:11.706 "data_size": 63488 00:11:11.706 }, 00:11:11.706 { 00:11:11.706 "name": "BaseBdev4", 00:11:11.706 "uuid": "5983d9fb-0f9a-56c5-b8b2-501f3593111c", 00:11:11.706 "is_configured": true, 00:11:11.706 "data_offset": 2048, 00:11:11.706 "data_size": 63488 00:11:11.706 } 00:11:11.706 ] 00:11:11.706 }' 00:11:11.706 15:17:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:11.706 15:17:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:12.275 15:17:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:12.275 15:17:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:12.275 15:17:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:12.275 [2024-11-19 15:17:02.335223] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:12.275 [2024-11-19 15:17:02.335351] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:12.275 [2024-11-19 15:17:02.337903] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:12.275 [2024-11-19 15:17:02.337962] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:12.275 [2024-11-19 15:17:02.338030] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:12.275 [2024-11-19 15:17:02.338049] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state offline 00:11:12.275 15:17:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:12.275 { 00:11:12.275 "results": [ 00:11:12.275 { 00:11:12.275 "job": "raid_bdev1", 00:11:12.275 "core_mask": "0x1", 00:11:12.275 "workload": "randrw", 00:11:12.275 "percentage": 50, 00:11:12.275 "status": "finished", 00:11:12.275 "queue_depth": 1, 00:11:12.275 "io_size": 131072, 00:11:12.275 "runtime": 1.434099, 00:11:12.275 "iops": 14271.678594016174, 00:11:12.275 "mibps": 1783.9598242520217, 00:11:12.275 "io_failed": 1, 00:11:12.275 "io_timeout": 0, 00:11:12.275 "avg_latency_us": 98.61633923397734, 00:11:12.275 "min_latency_us": 24.482096069868994, 00:11:12.275 "max_latency_us": 1359.3711790393013 00:11:12.275 } 00:11:12.275 ], 00:11:12.275 "core_count": 1 00:11:12.275 } 00:11:12.275 15:17:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 81918 00:11:12.275 15:17:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 81918 ']' 00:11:12.275 15:17:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 81918 00:11:12.275 15:17:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:11:12.275 15:17:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:11:12.275 15:17:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 81918 00:11:12.275 15:17:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:11:12.275 15:17:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:11:12.275 15:17:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 81918' 00:11:12.275 killing process with pid 81918 00:11:12.275 15:17:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 81918 00:11:12.275 [2024-11-19 15:17:02.383024] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:12.275 15:17:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 81918 00:11:12.275 [2024-11-19 15:17:02.448950] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:12.536 15:17:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.0aldQlrCxa 00:11:12.536 15:17:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:11:12.536 15:17:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:11:12.536 15:17:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.70 00:11:12.536 15:17:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:11:12.536 15:17:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:11:12.536 15:17:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:11:12.536 15:17:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.70 != \0\.\0\0 ]] 00:11:12.536 00:11:12.536 real 0m3.529s 00:11:12.536 user 0m4.339s 00:11:12.536 sys 0m0.626s 00:11:12.536 ************************************ 00:11:12.536 END TEST raid_read_error_test 00:11:12.536 ************************************ 00:11:12.536 15:17:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:12.536 15:17:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:12.536 15:17:02 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid0 4 write 00:11:12.536 15:17:02 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:11:12.536 15:17:02 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:12.536 15:17:02 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:12.536 ************************************ 00:11:12.536 START TEST raid_write_error_test 00:11:12.536 ************************************ 00:11:12.536 15:17:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid0 4 write 00:11:12.536 15:17:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:11:12.536 15:17:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:11:12.536 15:17:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:11:12.536 15:17:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:11:12.536 15:17:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:12.536 15:17:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:11:12.536 15:17:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:12.536 15:17:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:12.536 15:17:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:11:12.536 15:17:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:12.536 15:17:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:12.536 15:17:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:11:12.536 15:17:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:12.536 15:17:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:12.536 15:17:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:11:12.536 15:17:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:12.536 15:17:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:12.536 15:17:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:11:12.536 15:17:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:11:12.536 15:17:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:11:12.536 15:17:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:11:12.536 15:17:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:11:12.536 15:17:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:11:12.536 15:17:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:11:12.536 15:17:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:11:12.536 15:17:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:11:12.536 15:17:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:11:12.536 15:17:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:11:12.536 15:17:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.zV0eKYYjyR 00:11:12.796 15:17:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=82048 00:11:12.796 15:17:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:11:12.796 15:17:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 82048 00:11:12.796 15:17:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 82048 ']' 00:11:12.796 15:17:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:12.796 15:17:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:12.796 15:17:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:12.796 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:12.796 15:17:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:12.796 15:17:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:12.796 [2024-11-19 15:17:02.953986] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:11:12.796 [2024-11-19 15:17:02.954203] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid82048 ] 00:11:12.796 [2024-11-19 15:17:03.107355] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:13.056 [2024-11-19 15:17:03.146510] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:13.056 [2024-11-19 15:17:03.222817] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:13.056 [2024-11-19 15:17:03.222947] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:13.631 15:17:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:13.631 15:17:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:11:13.631 15:17:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:13.631 15:17:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:11:13.631 15:17:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:13.631 15:17:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:13.631 BaseBdev1_malloc 00:11:13.631 15:17:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:13.631 15:17:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:11:13.631 15:17:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:13.631 15:17:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:13.631 true 00:11:13.631 15:17:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:13.631 15:17:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:11:13.631 15:17:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:13.631 15:17:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:13.631 [2024-11-19 15:17:03.825176] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:11:13.631 [2024-11-19 15:17:03.825331] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:13.631 [2024-11-19 15:17:03.825364] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:11:13.631 [2024-11-19 15:17:03.825382] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:13.631 [2024-11-19 15:17:03.827946] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:13.631 [2024-11-19 15:17:03.827993] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:13.631 BaseBdev1 00:11:13.631 15:17:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:13.631 15:17:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:13.631 15:17:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:11:13.631 15:17:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:13.631 15:17:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:13.631 BaseBdev2_malloc 00:11:13.631 15:17:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:13.631 15:17:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:11:13.631 15:17:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:13.631 15:17:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:13.631 true 00:11:13.631 15:17:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:13.631 15:17:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:11:13.631 15:17:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:13.631 15:17:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:13.631 [2024-11-19 15:17:03.871972] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:11:13.631 [2024-11-19 15:17:03.872106] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:13.631 [2024-11-19 15:17:03.872130] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:11:13.631 [2024-11-19 15:17:03.872149] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:13.631 [2024-11-19 15:17:03.874553] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:13.631 [2024-11-19 15:17:03.874592] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:11:13.631 BaseBdev2 00:11:13.631 15:17:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:13.631 15:17:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:13.631 15:17:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:11:13.631 15:17:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:13.631 15:17:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:13.631 BaseBdev3_malloc 00:11:13.631 15:17:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:13.631 15:17:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:11:13.631 15:17:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:13.631 15:17:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:13.631 true 00:11:13.631 15:17:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:13.631 15:17:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:11:13.631 15:17:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:13.631 15:17:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:13.632 [2024-11-19 15:17:03.918577] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:11:13.632 [2024-11-19 15:17:03.918626] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:13.632 [2024-11-19 15:17:03.918644] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:11:13.632 [2024-11-19 15:17:03.918653] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:13.632 [2024-11-19 15:17:03.921096] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:13.632 [2024-11-19 15:17:03.921212] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:11:13.632 BaseBdev3 00:11:13.632 15:17:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:13.632 15:17:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:13.632 15:17:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:11:13.632 15:17:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:13.632 15:17:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:13.632 BaseBdev4_malloc 00:11:13.632 15:17:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:13.632 15:17:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:11:13.632 15:17:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:13.632 15:17:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:13.905 true 00:11:13.905 15:17:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:13.905 15:17:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:11:13.905 15:17:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:13.905 15:17:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:13.905 [2024-11-19 15:17:03.979447] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:11:13.905 [2024-11-19 15:17:03.979503] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:13.905 [2024-11-19 15:17:03.979531] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:11:13.905 [2024-11-19 15:17:03.979541] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:13.905 [2024-11-19 15:17:03.982085] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:13.905 [2024-11-19 15:17:03.982120] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:11:13.905 BaseBdev4 00:11:13.905 15:17:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:13.905 15:17:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:11:13.905 15:17:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:13.905 15:17:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:13.905 [2024-11-19 15:17:03.991486] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:13.905 [2024-11-19 15:17:03.993800] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:13.905 [2024-11-19 15:17:03.993982] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:13.905 [2024-11-19 15:17:03.994050] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:13.905 [2024-11-19 15:17:03.994265] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002000 00:11:13.905 [2024-11-19 15:17:03.994278] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:11:13.906 [2024-11-19 15:17:03.994557] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002ef0 00:11:13.906 [2024-11-19 15:17:03.994700] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002000 00:11:13.906 [2024-11-19 15:17:03.994714] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002000 00:11:13.906 [2024-11-19 15:17:03.994846] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:13.906 15:17:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:13.906 15:17:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:11:13.906 15:17:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:13.906 15:17:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:13.906 15:17:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:11:13.906 15:17:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:13.906 15:17:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:13.906 15:17:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:13.906 15:17:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:13.906 15:17:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:13.906 15:17:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:13.906 15:17:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:13.906 15:17:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:13.906 15:17:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:13.906 15:17:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:13.906 15:17:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:13.906 15:17:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:13.906 "name": "raid_bdev1", 00:11:13.906 "uuid": "c6017748-e9a8-45d1-a5a8-1ca2167f6fb3", 00:11:13.906 "strip_size_kb": 64, 00:11:13.906 "state": "online", 00:11:13.906 "raid_level": "raid0", 00:11:13.906 "superblock": true, 00:11:13.906 "num_base_bdevs": 4, 00:11:13.906 "num_base_bdevs_discovered": 4, 00:11:13.906 "num_base_bdevs_operational": 4, 00:11:13.906 "base_bdevs_list": [ 00:11:13.906 { 00:11:13.906 "name": "BaseBdev1", 00:11:13.906 "uuid": "64e22a47-fedd-547f-9a5a-feddaa3fabad", 00:11:13.906 "is_configured": true, 00:11:13.906 "data_offset": 2048, 00:11:13.906 "data_size": 63488 00:11:13.906 }, 00:11:13.906 { 00:11:13.906 "name": "BaseBdev2", 00:11:13.906 "uuid": "f010d629-0bed-533c-9b10-2e68a76c7c33", 00:11:13.906 "is_configured": true, 00:11:13.906 "data_offset": 2048, 00:11:13.906 "data_size": 63488 00:11:13.906 }, 00:11:13.906 { 00:11:13.906 "name": "BaseBdev3", 00:11:13.906 "uuid": "ee8b7f43-c7c1-5985-b432-8960debb27a8", 00:11:13.906 "is_configured": true, 00:11:13.906 "data_offset": 2048, 00:11:13.906 "data_size": 63488 00:11:13.906 }, 00:11:13.906 { 00:11:13.906 "name": "BaseBdev4", 00:11:13.906 "uuid": "7cc011ae-4dfe-5f52-baa0-61263f4603b9", 00:11:13.906 "is_configured": true, 00:11:13.906 "data_offset": 2048, 00:11:13.906 "data_size": 63488 00:11:13.906 } 00:11:13.906 ] 00:11:13.906 }' 00:11:13.906 15:17:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:13.906 15:17:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:14.166 15:17:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:11:14.166 15:17:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:11:14.425 [2024-11-19 15:17:04.523136] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000003090 00:11:15.365 15:17:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:11:15.365 15:17:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:15.365 15:17:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:15.365 15:17:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:15.365 15:17:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:11:15.365 15:17:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:11:15.365 15:17:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:11:15.365 15:17:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:11:15.365 15:17:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:15.365 15:17:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:15.365 15:17:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:11:15.365 15:17:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:15.365 15:17:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:15.365 15:17:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:15.365 15:17:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:15.365 15:17:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:15.365 15:17:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:15.365 15:17:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:15.365 15:17:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:15.365 15:17:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:15.365 15:17:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:15.365 15:17:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:15.365 15:17:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:15.365 "name": "raid_bdev1", 00:11:15.365 "uuid": "c6017748-e9a8-45d1-a5a8-1ca2167f6fb3", 00:11:15.365 "strip_size_kb": 64, 00:11:15.365 "state": "online", 00:11:15.365 "raid_level": "raid0", 00:11:15.365 "superblock": true, 00:11:15.365 "num_base_bdevs": 4, 00:11:15.365 "num_base_bdevs_discovered": 4, 00:11:15.365 "num_base_bdevs_operational": 4, 00:11:15.365 "base_bdevs_list": [ 00:11:15.365 { 00:11:15.365 "name": "BaseBdev1", 00:11:15.365 "uuid": "64e22a47-fedd-547f-9a5a-feddaa3fabad", 00:11:15.365 "is_configured": true, 00:11:15.365 "data_offset": 2048, 00:11:15.365 "data_size": 63488 00:11:15.365 }, 00:11:15.365 { 00:11:15.365 "name": "BaseBdev2", 00:11:15.365 "uuid": "f010d629-0bed-533c-9b10-2e68a76c7c33", 00:11:15.365 "is_configured": true, 00:11:15.365 "data_offset": 2048, 00:11:15.365 "data_size": 63488 00:11:15.365 }, 00:11:15.365 { 00:11:15.365 "name": "BaseBdev3", 00:11:15.365 "uuid": "ee8b7f43-c7c1-5985-b432-8960debb27a8", 00:11:15.365 "is_configured": true, 00:11:15.365 "data_offset": 2048, 00:11:15.365 "data_size": 63488 00:11:15.365 }, 00:11:15.365 { 00:11:15.365 "name": "BaseBdev4", 00:11:15.365 "uuid": "7cc011ae-4dfe-5f52-baa0-61263f4603b9", 00:11:15.365 "is_configured": true, 00:11:15.365 "data_offset": 2048, 00:11:15.365 "data_size": 63488 00:11:15.365 } 00:11:15.365 ] 00:11:15.365 }' 00:11:15.365 15:17:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:15.365 15:17:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:15.625 15:17:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:15.625 15:17:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:15.625 15:17:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:15.625 [2024-11-19 15:17:05.915900] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:15.625 [2024-11-19 15:17:05.915948] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:15.625 [2024-11-19 15:17:05.918384] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:15.625 [2024-11-19 15:17:05.918460] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:15.625 [2024-11-19 15:17:05.918515] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:15.626 [2024-11-19 15:17:05.918541] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state offline 00:11:15.626 { 00:11:15.626 "results": [ 00:11:15.626 { 00:11:15.626 "job": "raid_bdev1", 00:11:15.626 "core_mask": "0x1", 00:11:15.626 "workload": "randrw", 00:11:15.626 "percentage": 50, 00:11:15.626 "status": "finished", 00:11:15.626 "queue_depth": 1, 00:11:15.626 "io_size": 131072, 00:11:15.626 "runtime": 1.393263, 00:11:15.626 "iops": 14191.86470896019, 00:11:15.626 "mibps": 1773.9830886200236, 00:11:15.626 "io_failed": 1, 00:11:15.626 "io_timeout": 0, 00:11:15.626 "avg_latency_us": 99.21760611062209, 00:11:15.626 "min_latency_us": 24.929257641921396, 00:11:15.626 "max_latency_us": 1423.7624454148472 00:11:15.626 } 00:11:15.626 ], 00:11:15.626 "core_count": 1 00:11:15.626 } 00:11:15.626 15:17:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:15.626 15:17:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 82048 00:11:15.626 15:17:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 82048 ']' 00:11:15.626 15:17:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 82048 00:11:15.626 15:17:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:11:15.626 15:17:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:11:15.626 15:17:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 82048 00:11:15.886 killing process with pid 82048 00:11:15.886 15:17:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:11:15.886 15:17:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:11:15.886 15:17:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 82048' 00:11:15.886 15:17:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 82048 00:11:15.886 [2024-11-19 15:17:05.967493] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:15.886 15:17:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 82048 00:11:15.886 [2024-11-19 15:17:06.033968] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:16.146 15:17:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.zV0eKYYjyR 00:11:16.146 15:17:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:11:16.146 15:17:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:11:16.146 15:17:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.72 00:11:16.146 15:17:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:11:16.146 15:17:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:11:16.146 15:17:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:11:16.146 15:17:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.72 != \0\.\0\0 ]] 00:11:16.146 00:11:16.146 real 0m3.515s 00:11:16.146 user 0m4.292s 00:11:16.146 sys 0m0.626s 00:11:16.146 15:17:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:16.146 15:17:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:16.146 ************************************ 00:11:16.146 END TEST raid_write_error_test 00:11:16.146 ************************************ 00:11:16.146 15:17:06 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:11:16.146 15:17:06 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test concat 4 false 00:11:16.146 15:17:06 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:11:16.146 15:17:06 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:16.146 15:17:06 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:16.146 ************************************ 00:11:16.146 START TEST raid_state_function_test 00:11:16.146 ************************************ 00:11:16.146 15:17:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test concat 4 false 00:11:16.146 15:17:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:11:16.146 15:17:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:11:16.146 15:17:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:11:16.146 15:17:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:11:16.146 15:17:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:11:16.146 15:17:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:16.146 15:17:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:11:16.146 15:17:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:11:16.146 15:17:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:16.146 15:17:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:11:16.146 15:17:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:11:16.146 15:17:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:16.146 15:17:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:11:16.146 15:17:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:11:16.146 15:17:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:16.146 15:17:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:11:16.146 15:17:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:11:16.146 15:17:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:16.146 15:17:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:11:16.146 15:17:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:11:16.146 15:17:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:11:16.146 15:17:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:11:16.146 15:17:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:11:16.146 15:17:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:11:16.146 15:17:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:11:16.146 15:17:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:11:16.146 15:17:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:11:16.146 15:17:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:11:16.146 15:17:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:11:16.146 15:17:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=82181 00:11:16.146 15:17:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:11:16.146 15:17:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 82181' 00:11:16.146 Process raid pid: 82181 00:11:16.146 15:17:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 82181 00:11:16.146 15:17:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 82181 ']' 00:11:16.146 15:17:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:16.146 15:17:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:16.146 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:16.146 15:17:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:16.146 15:17:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:16.146 15:17:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:16.406 [2024-11-19 15:17:06.534346] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:11:16.406 [2024-11-19 15:17:06.534484] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:11:16.406 [2024-11-19 15:17:06.688301] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:16.406 [2024-11-19 15:17:06.728845] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:16.666 [2024-11-19 15:17:06.805478] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:16.666 [2024-11-19 15:17:06.805536] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:17.235 15:17:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:17.235 15:17:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:11:17.235 15:17:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:17.235 15:17:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:17.235 15:17:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:17.235 [2024-11-19 15:17:07.376811] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:11:17.235 [2024-11-19 15:17:07.376896] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:11:17.235 [2024-11-19 15:17:07.376907] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:17.235 [2024-11-19 15:17:07.376918] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:17.235 [2024-11-19 15:17:07.376924] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:11:17.235 [2024-11-19 15:17:07.376936] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:11:17.235 [2024-11-19 15:17:07.376942] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:11:17.235 [2024-11-19 15:17:07.376951] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:11:17.235 15:17:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:17.235 15:17:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:17.235 15:17:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:17.235 15:17:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:17.235 15:17:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:17.235 15:17:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:17.235 15:17:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:17.235 15:17:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:17.235 15:17:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:17.235 15:17:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:17.235 15:17:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:17.235 15:17:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:17.235 15:17:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:17.235 15:17:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:17.235 15:17:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:17.235 15:17:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:17.235 15:17:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:17.235 "name": "Existed_Raid", 00:11:17.235 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:17.235 "strip_size_kb": 64, 00:11:17.235 "state": "configuring", 00:11:17.235 "raid_level": "concat", 00:11:17.235 "superblock": false, 00:11:17.235 "num_base_bdevs": 4, 00:11:17.235 "num_base_bdevs_discovered": 0, 00:11:17.235 "num_base_bdevs_operational": 4, 00:11:17.235 "base_bdevs_list": [ 00:11:17.235 { 00:11:17.235 "name": "BaseBdev1", 00:11:17.235 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:17.235 "is_configured": false, 00:11:17.235 "data_offset": 0, 00:11:17.235 "data_size": 0 00:11:17.235 }, 00:11:17.235 { 00:11:17.235 "name": "BaseBdev2", 00:11:17.235 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:17.235 "is_configured": false, 00:11:17.235 "data_offset": 0, 00:11:17.235 "data_size": 0 00:11:17.235 }, 00:11:17.235 { 00:11:17.235 "name": "BaseBdev3", 00:11:17.235 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:17.235 "is_configured": false, 00:11:17.235 "data_offset": 0, 00:11:17.235 "data_size": 0 00:11:17.235 }, 00:11:17.235 { 00:11:17.235 "name": "BaseBdev4", 00:11:17.235 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:17.235 "is_configured": false, 00:11:17.235 "data_offset": 0, 00:11:17.235 "data_size": 0 00:11:17.235 } 00:11:17.235 ] 00:11:17.235 }' 00:11:17.235 15:17:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:17.235 15:17:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:17.495 15:17:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:11:17.495 15:17:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:17.495 15:17:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:17.495 [2024-11-19 15:17:07.800029] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:11:17.495 [2024-11-19 15:17:07.800082] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:11:17.495 15:17:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:17.495 15:17:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:17.495 15:17:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:17.495 15:17:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:17.495 [2024-11-19 15:17:07.812011] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:11:17.495 [2024-11-19 15:17:07.812057] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:11:17.495 [2024-11-19 15:17:07.812067] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:17.495 [2024-11-19 15:17:07.812077] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:17.495 [2024-11-19 15:17:07.812083] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:11:17.495 [2024-11-19 15:17:07.812093] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:11:17.495 [2024-11-19 15:17:07.812098] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:11:17.495 [2024-11-19 15:17:07.812109] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:11:17.495 15:17:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:17.495 15:17:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:11:17.496 15:17:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:17.496 15:17:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:17.756 [2024-11-19 15:17:07.839165] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:17.756 BaseBdev1 00:11:17.756 15:17:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:17.756 15:17:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:11:17.756 15:17:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:11:17.756 15:17:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:17.756 15:17:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:11:17.756 15:17:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:17.756 15:17:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:17.756 15:17:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:17.756 15:17:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:17.756 15:17:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:17.756 15:17:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:17.756 15:17:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:11:17.756 15:17:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:17.756 15:17:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:17.756 [ 00:11:17.756 { 00:11:17.756 "name": "BaseBdev1", 00:11:17.756 "aliases": [ 00:11:17.756 "59489e39-639e-4307-bebc-f566e8c7868d" 00:11:17.756 ], 00:11:17.756 "product_name": "Malloc disk", 00:11:17.756 "block_size": 512, 00:11:17.756 "num_blocks": 65536, 00:11:17.756 "uuid": "59489e39-639e-4307-bebc-f566e8c7868d", 00:11:17.756 "assigned_rate_limits": { 00:11:17.756 "rw_ios_per_sec": 0, 00:11:17.756 "rw_mbytes_per_sec": 0, 00:11:17.756 "r_mbytes_per_sec": 0, 00:11:17.756 "w_mbytes_per_sec": 0 00:11:17.756 }, 00:11:17.756 "claimed": true, 00:11:17.756 "claim_type": "exclusive_write", 00:11:17.756 "zoned": false, 00:11:17.756 "supported_io_types": { 00:11:17.756 "read": true, 00:11:17.756 "write": true, 00:11:17.756 "unmap": true, 00:11:17.756 "flush": true, 00:11:17.756 "reset": true, 00:11:17.756 "nvme_admin": false, 00:11:17.756 "nvme_io": false, 00:11:17.756 "nvme_io_md": false, 00:11:17.756 "write_zeroes": true, 00:11:17.756 "zcopy": true, 00:11:17.756 "get_zone_info": false, 00:11:17.756 "zone_management": false, 00:11:17.756 "zone_append": false, 00:11:17.756 "compare": false, 00:11:17.756 "compare_and_write": false, 00:11:17.756 "abort": true, 00:11:17.756 "seek_hole": false, 00:11:17.756 "seek_data": false, 00:11:17.756 "copy": true, 00:11:17.756 "nvme_iov_md": false 00:11:17.756 }, 00:11:17.756 "memory_domains": [ 00:11:17.756 { 00:11:17.756 "dma_device_id": "system", 00:11:17.756 "dma_device_type": 1 00:11:17.756 }, 00:11:17.756 { 00:11:17.756 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:17.756 "dma_device_type": 2 00:11:17.756 } 00:11:17.756 ], 00:11:17.756 "driver_specific": {} 00:11:17.756 } 00:11:17.756 ] 00:11:17.756 15:17:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:17.756 15:17:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:11:17.756 15:17:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:17.756 15:17:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:17.756 15:17:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:17.756 15:17:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:17.756 15:17:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:17.756 15:17:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:17.756 15:17:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:17.756 15:17:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:17.756 15:17:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:17.756 15:17:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:17.756 15:17:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:17.756 15:17:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:17.756 15:17:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:17.756 15:17:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:17.756 15:17:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:17.756 15:17:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:17.756 "name": "Existed_Raid", 00:11:17.756 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:17.756 "strip_size_kb": 64, 00:11:17.756 "state": "configuring", 00:11:17.756 "raid_level": "concat", 00:11:17.756 "superblock": false, 00:11:17.756 "num_base_bdevs": 4, 00:11:17.756 "num_base_bdevs_discovered": 1, 00:11:17.756 "num_base_bdevs_operational": 4, 00:11:17.756 "base_bdevs_list": [ 00:11:17.757 { 00:11:17.757 "name": "BaseBdev1", 00:11:17.757 "uuid": "59489e39-639e-4307-bebc-f566e8c7868d", 00:11:17.757 "is_configured": true, 00:11:17.757 "data_offset": 0, 00:11:17.757 "data_size": 65536 00:11:17.757 }, 00:11:17.757 { 00:11:17.757 "name": "BaseBdev2", 00:11:17.757 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:17.757 "is_configured": false, 00:11:17.757 "data_offset": 0, 00:11:17.757 "data_size": 0 00:11:17.757 }, 00:11:17.757 { 00:11:17.757 "name": "BaseBdev3", 00:11:17.757 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:17.757 "is_configured": false, 00:11:17.757 "data_offset": 0, 00:11:17.757 "data_size": 0 00:11:17.757 }, 00:11:17.757 { 00:11:17.757 "name": "BaseBdev4", 00:11:17.757 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:17.757 "is_configured": false, 00:11:17.757 "data_offset": 0, 00:11:17.757 "data_size": 0 00:11:17.757 } 00:11:17.757 ] 00:11:17.757 }' 00:11:17.757 15:17:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:17.757 15:17:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:18.016 15:17:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:11:18.016 15:17:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:18.016 15:17:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:18.016 [2024-11-19 15:17:08.330396] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:11:18.016 [2024-11-19 15:17:08.330470] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:11:18.016 15:17:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:18.016 15:17:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:18.016 15:17:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:18.016 15:17:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:18.016 [2024-11-19 15:17:08.338409] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:18.016 [2024-11-19 15:17:08.340584] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:18.016 [2024-11-19 15:17:08.340631] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:18.017 [2024-11-19 15:17:08.340641] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:11:18.017 [2024-11-19 15:17:08.340650] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:11:18.017 [2024-11-19 15:17:08.340657] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:11:18.017 [2024-11-19 15:17:08.340665] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:11:18.017 15:17:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:18.017 15:17:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:11:18.017 15:17:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:18.017 15:17:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:18.017 15:17:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:18.017 15:17:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:18.017 15:17:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:18.017 15:17:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:18.017 15:17:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:18.017 15:17:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:18.017 15:17:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:18.017 15:17:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:18.017 15:17:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:18.017 15:17:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:18.017 15:17:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:18.017 15:17:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:18.017 15:17:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:18.290 15:17:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:18.290 15:17:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:18.290 "name": "Existed_Raid", 00:11:18.290 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:18.290 "strip_size_kb": 64, 00:11:18.290 "state": "configuring", 00:11:18.290 "raid_level": "concat", 00:11:18.290 "superblock": false, 00:11:18.291 "num_base_bdevs": 4, 00:11:18.291 "num_base_bdevs_discovered": 1, 00:11:18.291 "num_base_bdevs_operational": 4, 00:11:18.291 "base_bdevs_list": [ 00:11:18.291 { 00:11:18.291 "name": "BaseBdev1", 00:11:18.291 "uuid": "59489e39-639e-4307-bebc-f566e8c7868d", 00:11:18.291 "is_configured": true, 00:11:18.291 "data_offset": 0, 00:11:18.291 "data_size": 65536 00:11:18.291 }, 00:11:18.291 { 00:11:18.291 "name": "BaseBdev2", 00:11:18.291 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:18.291 "is_configured": false, 00:11:18.291 "data_offset": 0, 00:11:18.291 "data_size": 0 00:11:18.291 }, 00:11:18.291 { 00:11:18.291 "name": "BaseBdev3", 00:11:18.291 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:18.291 "is_configured": false, 00:11:18.291 "data_offset": 0, 00:11:18.291 "data_size": 0 00:11:18.291 }, 00:11:18.291 { 00:11:18.291 "name": "BaseBdev4", 00:11:18.291 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:18.291 "is_configured": false, 00:11:18.291 "data_offset": 0, 00:11:18.291 "data_size": 0 00:11:18.291 } 00:11:18.291 ] 00:11:18.291 }' 00:11:18.291 15:17:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:18.291 15:17:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:18.551 15:17:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:11:18.551 15:17:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:18.551 15:17:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:18.551 [2024-11-19 15:17:08.770418] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:18.551 BaseBdev2 00:11:18.551 15:17:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:18.551 15:17:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:11:18.551 15:17:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:11:18.551 15:17:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:18.551 15:17:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:11:18.551 15:17:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:18.551 15:17:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:18.551 15:17:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:18.551 15:17:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:18.551 15:17:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:18.551 15:17:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:18.551 15:17:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:11:18.551 15:17:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:18.551 15:17:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:18.551 [ 00:11:18.551 { 00:11:18.551 "name": "BaseBdev2", 00:11:18.551 "aliases": [ 00:11:18.551 "1bb986de-34c6-47c3-a13f-782ae8d0b192" 00:11:18.551 ], 00:11:18.551 "product_name": "Malloc disk", 00:11:18.551 "block_size": 512, 00:11:18.551 "num_blocks": 65536, 00:11:18.551 "uuid": "1bb986de-34c6-47c3-a13f-782ae8d0b192", 00:11:18.551 "assigned_rate_limits": { 00:11:18.551 "rw_ios_per_sec": 0, 00:11:18.551 "rw_mbytes_per_sec": 0, 00:11:18.551 "r_mbytes_per_sec": 0, 00:11:18.551 "w_mbytes_per_sec": 0 00:11:18.551 }, 00:11:18.551 "claimed": true, 00:11:18.551 "claim_type": "exclusive_write", 00:11:18.551 "zoned": false, 00:11:18.551 "supported_io_types": { 00:11:18.551 "read": true, 00:11:18.551 "write": true, 00:11:18.551 "unmap": true, 00:11:18.551 "flush": true, 00:11:18.551 "reset": true, 00:11:18.551 "nvme_admin": false, 00:11:18.551 "nvme_io": false, 00:11:18.551 "nvme_io_md": false, 00:11:18.551 "write_zeroes": true, 00:11:18.551 "zcopy": true, 00:11:18.551 "get_zone_info": false, 00:11:18.551 "zone_management": false, 00:11:18.551 "zone_append": false, 00:11:18.551 "compare": false, 00:11:18.551 "compare_and_write": false, 00:11:18.551 "abort": true, 00:11:18.551 "seek_hole": false, 00:11:18.551 "seek_data": false, 00:11:18.551 "copy": true, 00:11:18.551 "nvme_iov_md": false 00:11:18.551 }, 00:11:18.551 "memory_domains": [ 00:11:18.551 { 00:11:18.551 "dma_device_id": "system", 00:11:18.551 "dma_device_type": 1 00:11:18.551 }, 00:11:18.551 { 00:11:18.551 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:18.551 "dma_device_type": 2 00:11:18.551 } 00:11:18.551 ], 00:11:18.551 "driver_specific": {} 00:11:18.551 } 00:11:18.551 ] 00:11:18.551 15:17:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:18.551 15:17:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:11:18.551 15:17:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:11:18.551 15:17:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:18.551 15:17:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:18.551 15:17:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:18.551 15:17:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:18.551 15:17:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:18.551 15:17:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:18.551 15:17:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:18.551 15:17:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:18.551 15:17:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:18.551 15:17:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:18.551 15:17:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:18.551 15:17:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:18.551 15:17:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:18.551 15:17:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:18.551 15:17:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:18.551 15:17:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:18.551 15:17:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:18.551 "name": "Existed_Raid", 00:11:18.551 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:18.551 "strip_size_kb": 64, 00:11:18.551 "state": "configuring", 00:11:18.551 "raid_level": "concat", 00:11:18.551 "superblock": false, 00:11:18.551 "num_base_bdevs": 4, 00:11:18.551 "num_base_bdevs_discovered": 2, 00:11:18.551 "num_base_bdevs_operational": 4, 00:11:18.551 "base_bdevs_list": [ 00:11:18.551 { 00:11:18.551 "name": "BaseBdev1", 00:11:18.551 "uuid": "59489e39-639e-4307-bebc-f566e8c7868d", 00:11:18.551 "is_configured": true, 00:11:18.551 "data_offset": 0, 00:11:18.551 "data_size": 65536 00:11:18.551 }, 00:11:18.551 { 00:11:18.551 "name": "BaseBdev2", 00:11:18.551 "uuid": "1bb986de-34c6-47c3-a13f-782ae8d0b192", 00:11:18.551 "is_configured": true, 00:11:18.551 "data_offset": 0, 00:11:18.551 "data_size": 65536 00:11:18.551 }, 00:11:18.551 { 00:11:18.551 "name": "BaseBdev3", 00:11:18.551 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:18.551 "is_configured": false, 00:11:18.551 "data_offset": 0, 00:11:18.551 "data_size": 0 00:11:18.551 }, 00:11:18.551 { 00:11:18.551 "name": "BaseBdev4", 00:11:18.551 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:18.551 "is_configured": false, 00:11:18.551 "data_offset": 0, 00:11:18.551 "data_size": 0 00:11:18.551 } 00:11:18.551 ] 00:11:18.551 }' 00:11:18.551 15:17:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:18.551 15:17:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:19.120 15:17:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:11:19.120 15:17:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:19.120 15:17:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:19.120 [2024-11-19 15:17:09.260615] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:19.120 BaseBdev3 00:11:19.120 15:17:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:19.120 15:17:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:11:19.120 15:17:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:11:19.120 15:17:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:19.120 15:17:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:11:19.120 15:17:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:19.120 15:17:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:19.120 15:17:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:19.120 15:17:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:19.120 15:17:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:19.120 15:17:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:19.120 15:17:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:11:19.120 15:17:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:19.120 15:17:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:19.120 [ 00:11:19.120 { 00:11:19.120 "name": "BaseBdev3", 00:11:19.120 "aliases": [ 00:11:19.120 "ba7c69dd-52c7-4771-8573-759b3ae8fa4c" 00:11:19.120 ], 00:11:19.120 "product_name": "Malloc disk", 00:11:19.120 "block_size": 512, 00:11:19.120 "num_blocks": 65536, 00:11:19.120 "uuid": "ba7c69dd-52c7-4771-8573-759b3ae8fa4c", 00:11:19.120 "assigned_rate_limits": { 00:11:19.120 "rw_ios_per_sec": 0, 00:11:19.120 "rw_mbytes_per_sec": 0, 00:11:19.120 "r_mbytes_per_sec": 0, 00:11:19.120 "w_mbytes_per_sec": 0 00:11:19.120 }, 00:11:19.120 "claimed": true, 00:11:19.120 "claim_type": "exclusive_write", 00:11:19.120 "zoned": false, 00:11:19.120 "supported_io_types": { 00:11:19.120 "read": true, 00:11:19.120 "write": true, 00:11:19.120 "unmap": true, 00:11:19.120 "flush": true, 00:11:19.120 "reset": true, 00:11:19.120 "nvme_admin": false, 00:11:19.120 "nvme_io": false, 00:11:19.120 "nvme_io_md": false, 00:11:19.120 "write_zeroes": true, 00:11:19.120 "zcopy": true, 00:11:19.120 "get_zone_info": false, 00:11:19.120 "zone_management": false, 00:11:19.120 "zone_append": false, 00:11:19.121 "compare": false, 00:11:19.121 "compare_and_write": false, 00:11:19.121 "abort": true, 00:11:19.121 "seek_hole": false, 00:11:19.121 "seek_data": false, 00:11:19.121 "copy": true, 00:11:19.121 "nvme_iov_md": false 00:11:19.121 }, 00:11:19.121 "memory_domains": [ 00:11:19.121 { 00:11:19.121 "dma_device_id": "system", 00:11:19.121 "dma_device_type": 1 00:11:19.121 }, 00:11:19.121 { 00:11:19.121 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:19.121 "dma_device_type": 2 00:11:19.121 } 00:11:19.121 ], 00:11:19.121 "driver_specific": {} 00:11:19.121 } 00:11:19.121 ] 00:11:19.121 15:17:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:19.121 15:17:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:11:19.121 15:17:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:11:19.121 15:17:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:19.121 15:17:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:19.121 15:17:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:19.121 15:17:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:19.121 15:17:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:19.121 15:17:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:19.121 15:17:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:19.121 15:17:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:19.121 15:17:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:19.121 15:17:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:19.121 15:17:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:19.121 15:17:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:19.121 15:17:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:19.121 15:17:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:19.121 15:17:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:19.121 15:17:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:19.121 15:17:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:19.121 "name": "Existed_Raid", 00:11:19.121 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:19.121 "strip_size_kb": 64, 00:11:19.121 "state": "configuring", 00:11:19.121 "raid_level": "concat", 00:11:19.121 "superblock": false, 00:11:19.121 "num_base_bdevs": 4, 00:11:19.121 "num_base_bdevs_discovered": 3, 00:11:19.121 "num_base_bdevs_operational": 4, 00:11:19.121 "base_bdevs_list": [ 00:11:19.121 { 00:11:19.121 "name": "BaseBdev1", 00:11:19.121 "uuid": "59489e39-639e-4307-bebc-f566e8c7868d", 00:11:19.121 "is_configured": true, 00:11:19.121 "data_offset": 0, 00:11:19.121 "data_size": 65536 00:11:19.121 }, 00:11:19.121 { 00:11:19.121 "name": "BaseBdev2", 00:11:19.121 "uuid": "1bb986de-34c6-47c3-a13f-782ae8d0b192", 00:11:19.121 "is_configured": true, 00:11:19.121 "data_offset": 0, 00:11:19.121 "data_size": 65536 00:11:19.121 }, 00:11:19.121 { 00:11:19.121 "name": "BaseBdev3", 00:11:19.121 "uuid": "ba7c69dd-52c7-4771-8573-759b3ae8fa4c", 00:11:19.121 "is_configured": true, 00:11:19.121 "data_offset": 0, 00:11:19.121 "data_size": 65536 00:11:19.121 }, 00:11:19.121 { 00:11:19.121 "name": "BaseBdev4", 00:11:19.121 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:19.121 "is_configured": false, 00:11:19.121 "data_offset": 0, 00:11:19.121 "data_size": 0 00:11:19.121 } 00:11:19.121 ] 00:11:19.121 }' 00:11:19.121 15:17:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:19.121 15:17:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:19.690 15:17:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:11:19.690 15:17:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:19.690 15:17:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:19.690 [2024-11-19 15:17:09.752811] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:19.690 [2024-11-19 15:17:09.752873] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:11:19.690 [2024-11-19 15:17:09.752882] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:11:19.690 [2024-11-19 15:17:09.753252] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:11:19.690 [2024-11-19 15:17:09.753417] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:11:19.690 [2024-11-19 15:17:09.753437] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:11:19.690 [2024-11-19 15:17:09.753660] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:19.690 BaseBdev4 00:11:19.690 15:17:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:19.690 15:17:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:11:19.690 15:17:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:11:19.690 15:17:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:19.690 15:17:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:11:19.690 15:17:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:19.690 15:17:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:19.690 15:17:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:19.690 15:17:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:19.690 15:17:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:19.690 15:17:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:19.690 15:17:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:11:19.691 15:17:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:19.691 15:17:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:19.691 [ 00:11:19.691 { 00:11:19.691 "name": "BaseBdev4", 00:11:19.691 "aliases": [ 00:11:19.691 "7ab2ccc7-ebea-4c2e-8341-87da8de7cca0" 00:11:19.691 ], 00:11:19.691 "product_name": "Malloc disk", 00:11:19.691 "block_size": 512, 00:11:19.691 "num_blocks": 65536, 00:11:19.691 "uuid": "7ab2ccc7-ebea-4c2e-8341-87da8de7cca0", 00:11:19.691 "assigned_rate_limits": { 00:11:19.691 "rw_ios_per_sec": 0, 00:11:19.691 "rw_mbytes_per_sec": 0, 00:11:19.691 "r_mbytes_per_sec": 0, 00:11:19.691 "w_mbytes_per_sec": 0 00:11:19.691 }, 00:11:19.691 "claimed": true, 00:11:19.691 "claim_type": "exclusive_write", 00:11:19.691 "zoned": false, 00:11:19.691 "supported_io_types": { 00:11:19.691 "read": true, 00:11:19.691 "write": true, 00:11:19.691 "unmap": true, 00:11:19.691 "flush": true, 00:11:19.691 "reset": true, 00:11:19.691 "nvme_admin": false, 00:11:19.691 "nvme_io": false, 00:11:19.691 "nvme_io_md": false, 00:11:19.691 "write_zeroes": true, 00:11:19.691 "zcopy": true, 00:11:19.691 "get_zone_info": false, 00:11:19.691 "zone_management": false, 00:11:19.691 "zone_append": false, 00:11:19.691 "compare": false, 00:11:19.691 "compare_and_write": false, 00:11:19.691 "abort": true, 00:11:19.691 "seek_hole": false, 00:11:19.691 "seek_data": false, 00:11:19.691 "copy": true, 00:11:19.691 "nvme_iov_md": false 00:11:19.691 }, 00:11:19.691 "memory_domains": [ 00:11:19.691 { 00:11:19.691 "dma_device_id": "system", 00:11:19.691 "dma_device_type": 1 00:11:19.691 }, 00:11:19.691 { 00:11:19.691 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:19.691 "dma_device_type": 2 00:11:19.691 } 00:11:19.691 ], 00:11:19.691 "driver_specific": {} 00:11:19.691 } 00:11:19.691 ] 00:11:19.691 15:17:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:19.691 15:17:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:11:19.691 15:17:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:11:19.691 15:17:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:19.691 15:17:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:11:19.691 15:17:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:19.691 15:17:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:19.691 15:17:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:19.691 15:17:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:19.691 15:17:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:19.691 15:17:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:19.691 15:17:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:19.691 15:17:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:19.691 15:17:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:19.691 15:17:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:19.691 15:17:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:19.691 15:17:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:19.691 15:17:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:19.691 15:17:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:19.691 15:17:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:19.691 "name": "Existed_Raid", 00:11:19.691 "uuid": "75fb8d0c-3b4e-46c2-8164-d7214b67ee69", 00:11:19.691 "strip_size_kb": 64, 00:11:19.691 "state": "online", 00:11:19.691 "raid_level": "concat", 00:11:19.691 "superblock": false, 00:11:19.691 "num_base_bdevs": 4, 00:11:19.691 "num_base_bdevs_discovered": 4, 00:11:19.691 "num_base_bdevs_operational": 4, 00:11:19.691 "base_bdevs_list": [ 00:11:19.691 { 00:11:19.691 "name": "BaseBdev1", 00:11:19.691 "uuid": "59489e39-639e-4307-bebc-f566e8c7868d", 00:11:19.691 "is_configured": true, 00:11:19.691 "data_offset": 0, 00:11:19.691 "data_size": 65536 00:11:19.691 }, 00:11:19.691 { 00:11:19.691 "name": "BaseBdev2", 00:11:19.691 "uuid": "1bb986de-34c6-47c3-a13f-782ae8d0b192", 00:11:19.691 "is_configured": true, 00:11:19.691 "data_offset": 0, 00:11:19.691 "data_size": 65536 00:11:19.691 }, 00:11:19.691 { 00:11:19.691 "name": "BaseBdev3", 00:11:19.691 "uuid": "ba7c69dd-52c7-4771-8573-759b3ae8fa4c", 00:11:19.691 "is_configured": true, 00:11:19.691 "data_offset": 0, 00:11:19.691 "data_size": 65536 00:11:19.691 }, 00:11:19.691 { 00:11:19.691 "name": "BaseBdev4", 00:11:19.691 "uuid": "7ab2ccc7-ebea-4c2e-8341-87da8de7cca0", 00:11:19.691 "is_configured": true, 00:11:19.691 "data_offset": 0, 00:11:19.691 "data_size": 65536 00:11:19.691 } 00:11:19.691 ] 00:11:19.691 }' 00:11:19.691 15:17:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:19.691 15:17:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:19.951 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:11:19.951 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:11:19.951 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:11:19.951 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:11:19.951 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:11:19.951 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:11:19.951 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:11:19.951 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:11:19.951 15:17:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:19.951 15:17:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:19.951 [2024-11-19 15:17:10.268289] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:20.211 15:17:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:20.211 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:11:20.211 "name": "Existed_Raid", 00:11:20.211 "aliases": [ 00:11:20.211 "75fb8d0c-3b4e-46c2-8164-d7214b67ee69" 00:11:20.211 ], 00:11:20.211 "product_name": "Raid Volume", 00:11:20.211 "block_size": 512, 00:11:20.211 "num_blocks": 262144, 00:11:20.211 "uuid": "75fb8d0c-3b4e-46c2-8164-d7214b67ee69", 00:11:20.211 "assigned_rate_limits": { 00:11:20.211 "rw_ios_per_sec": 0, 00:11:20.211 "rw_mbytes_per_sec": 0, 00:11:20.211 "r_mbytes_per_sec": 0, 00:11:20.211 "w_mbytes_per_sec": 0 00:11:20.211 }, 00:11:20.211 "claimed": false, 00:11:20.211 "zoned": false, 00:11:20.211 "supported_io_types": { 00:11:20.211 "read": true, 00:11:20.211 "write": true, 00:11:20.211 "unmap": true, 00:11:20.211 "flush": true, 00:11:20.211 "reset": true, 00:11:20.211 "nvme_admin": false, 00:11:20.211 "nvme_io": false, 00:11:20.211 "nvme_io_md": false, 00:11:20.211 "write_zeroes": true, 00:11:20.211 "zcopy": false, 00:11:20.211 "get_zone_info": false, 00:11:20.211 "zone_management": false, 00:11:20.211 "zone_append": false, 00:11:20.211 "compare": false, 00:11:20.211 "compare_and_write": false, 00:11:20.211 "abort": false, 00:11:20.211 "seek_hole": false, 00:11:20.211 "seek_data": false, 00:11:20.211 "copy": false, 00:11:20.211 "nvme_iov_md": false 00:11:20.211 }, 00:11:20.211 "memory_domains": [ 00:11:20.211 { 00:11:20.211 "dma_device_id": "system", 00:11:20.211 "dma_device_type": 1 00:11:20.211 }, 00:11:20.211 { 00:11:20.211 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:20.211 "dma_device_type": 2 00:11:20.211 }, 00:11:20.211 { 00:11:20.211 "dma_device_id": "system", 00:11:20.211 "dma_device_type": 1 00:11:20.211 }, 00:11:20.211 { 00:11:20.211 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:20.211 "dma_device_type": 2 00:11:20.211 }, 00:11:20.211 { 00:11:20.211 "dma_device_id": "system", 00:11:20.211 "dma_device_type": 1 00:11:20.211 }, 00:11:20.211 { 00:11:20.211 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:20.211 "dma_device_type": 2 00:11:20.211 }, 00:11:20.211 { 00:11:20.211 "dma_device_id": "system", 00:11:20.211 "dma_device_type": 1 00:11:20.211 }, 00:11:20.211 { 00:11:20.211 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:20.211 "dma_device_type": 2 00:11:20.211 } 00:11:20.211 ], 00:11:20.211 "driver_specific": { 00:11:20.211 "raid": { 00:11:20.211 "uuid": "75fb8d0c-3b4e-46c2-8164-d7214b67ee69", 00:11:20.211 "strip_size_kb": 64, 00:11:20.211 "state": "online", 00:11:20.211 "raid_level": "concat", 00:11:20.211 "superblock": false, 00:11:20.211 "num_base_bdevs": 4, 00:11:20.211 "num_base_bdevs_discovered": 4, 00:11:20.211 "num_base_bdevs_operational": 4, 00:11:20.211 "base_bdevs_list": [ 00:11:20.211 { 00:11:20.211 "name": "BaseBdev1", 00:11:20.211 "uuid": "59489e39-639e-4307-bebc-f566e8c7868d", 00:11:20.211 "is_configured": true, 00:11:20.211 "data_offset": 0, 00:11:20.211 "data_size": 65536 00:11:20.211 }, 00:11:20.211 { 00:11:20.211 "name": "BaseBdev2", 00:11:20.211 "uuid": "1bb986de-34c6-47c3-a13f-782ae8d0b192", 00:11:20.211 "is_configured": true, 00:11:20.211 "data_offset": 0, 00:11:20.211 "data_size": 65536 00:11:20.211 }, 00:11:20.211 { 00:11:20.211 "name": "BaseBdev3", 00:11:20.211 "uuid": "ba7c69dd-52c7-4771-8573-759b3ae8fa4c", 00:11:20.211 "is_configured": true, 00:11:20.211 "data_offset": 0, 00:11:20.211 "data_size": 65536 00:11:20.211 }, 00:11:20.211 { 00:11:20.211 "name": "BaseBdev4", 00:11:20.211 "uuid": "7ab2ccc7-ebea-4c2e-8341-87da8de7cca0", 00:11:20.211 "is_configured": true, 00:11:20.211 "data_offset": 0, 00:11:20.211 "data_size": 65536 00:11:20.211 } 00:11:20.211 ] 00:11:20.211 } 00:11:20.211 } 00:11:20.211 }' 00:11:20.211 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:11:20.211 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:11:20.211 BaseBdev2 00:11:20.211 BaseBdev3 00:11:20.211 BaseBdev4' 00:11:20.211 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:20.211 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:11:20.211 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:20.211 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:11:20.211 15:17:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:20.211 15:17:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.211 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:20.211 15:17:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:20.211 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:20.211 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:20.211 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:20.211 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:11:20.211 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:20.211 15:17:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:20.211 15:17:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.211 15:17:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:20.211 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:20.211 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:20.211 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:20.211 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:11:20.211 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:20.211 15:17:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:20.211 15:17:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.211 15:17:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:20.212 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:20.212 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:20.212 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:20.212 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:20.212 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:11:20.212 15:17:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:20.212 15:17:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.471 15:17:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:20.471 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:20.471 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:20.471 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:11:20.471 15:17:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:20.471 15:17:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.471 [2024-11-19 15:17:10.571582] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:20.471 [2024-11-19 15:17:10.571615] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:20.471 [2024-11-19 15:17:10.571679] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:20.471 15:17:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:20.471 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:11:20.471 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:11:20.471 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:11:20.472 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:11:20.472 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:11:20.472 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 3 00:11:20.472 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:20.472 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:11:20.472 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:20.472 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:20.472 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:11:20.472 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:20.472 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:20.472 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:20.472 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:20.472 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:20.472 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:20.472 15:17:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:20.472 15:17:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.472 15:17:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:20.472 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:20.472 "name": "Existed_Raid", 00:11:20.472 "uuid": "75fb8d0c-3b4e-46c2-8164-d7214b67ee69", 00:11:20.472 "strip_size_kb": 64, 00:11:20.472 "state": "offline", 00:11:20.472 "raid_level": "concat", 00:11:20.472 "superblock": false, 00:11:20.472 "num_base_bdevs": 4, 00:11:20.472 "num_base_bdevs_discovered": 3, 00:11:20.472 "num_base_bdevs_operational": 3, 00:11:20.472 "base_bdevs_list": [ 00:11:20.472 { 00:11:20.472 "name": null, 00:11:20.472 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:20.472 "is_configured": false, 00:11:20.472 "data_offset": 0, 00:11:20.472 "data_size": 65536 00:11:20.472 }, 00:11:20.472 { 00:11:20.472 "name": "BaseBdev2", 00:11:20.472 "uuid": "1bb986de-34c6-47c3-a13f-782ae8d0b192", 00:11:20.472 "is_configured": true, 00:11:20.472 "data_offset": 0, 00:11:20.472 "data_size": 65536 00:11:20.472 }, 00:11:20.472 { 00:11:20.472 "name": "BaseBdev3", 00:11:20.472 "uuid": "ba7c69dd-52c7-4771-8573-759b3ae8fa4c", 00:11:20.472 "is_configured": true, 00:11:20.472 "data_offset": 0, 00:11:20.472 "data_size": 65536 00:11:20.472 }, 00:11:20.472 { 00:11:20.472 "name": "BaseBdev4", 00:11:20.472 "uuid": "7ab2ccc7-ebea-4c2e-8341-87da8de7cca0", 00:11:20.472 "is_configured": true, 00:11:20.472 "data_offset": 0, 00:11:20.472 "data_size": 65536 00:11:20.472 } 00:11:20.472 ] 00:11:20.472 }' 00:11:20.472 15:17:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:20.472 15:17:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.731 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:11:20.731 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:11:20.731 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:20.731 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:11:20.731 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:20.731 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.731 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.023 [2024-11-19 15:17:11.091390] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.023 [2024-11-19 15:17:11.167343] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.023 [2024-11-19 15:17:11.246327] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:11:21.023 [2024-11-19 15:17:11.246373] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.023 BaseBdev2 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:21.023 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.283 [ 00:11:21.283 { 00:11:21.283 "name": "BaseBdev2", 00:11:21.283 "aliases": [ 00:11:21.283 "858748e1-8394-4f4e-a462-ecf04a6dd41c" 00:11:21.283 ], 00:11:21.283 "product_name": "Malloc disk", 00:11:21.283 "block_size": 512, 00:11:21.283 "num_blocks": 65536, 00:11:21.283 "uuid": "858748e1-8394-4f4e-a462-ecf04a6dd41c", 00:11:21.283 "assigned_rate_limits": { 00:11:21.283 "rw_ios_per_sec": 0, 00:11:21.283 "rw_mbytes_per_sec": 0, 00:11:21.283 "r_mbytes_per_sec": 0, 00:11:21.283 "w_mbytes_per_sec": 0 00:11:21.283 }, 00:11:21.283 "claimed": false, 00:11:21.283 "zoned": false, 00:11:21.283 "supported_io_types": { 00:11:21.283 "read": true, 00:11:21.283 "write": true, 00:11:21.283 "unmap": true, 00:11:21.283 "flush": true, 00:11:21.283 "reset": true, 00:11:21.283 "nvme_admin": false, 00:11:21.283 "nvme_io": false, 00:11:21.283 "nvme_io_md": false, 00:11:21.283 "write_zeroes": true, 00:11:21.283 "zcopy": true, 00:11:21.283 "get_zone_info": false, 00:11:21.283 "zone_management": false, 00:11:21.283 "zone_append": false, 00:11:21.283 "compare": false, 00:11:21.283 "compare_and_write": false, 00:11:21.283 "abort": true, 00:11:21.283 "seek_hole": false, 00:11:21.283 "seek_data": false, 00:11:21.283 "copy": true, 00:11:21.283 "nvme_iov_md": false 00:11:21.283 }, 00:11:21.283 "memory_domains": [ 00:11:21.283 { 00:11:21.283 "dma_device_id": "system", 00:11:21.283 "dma_device_type": 1 00:11:21.283 }, 00:11:21.283 { 00:11:21.283 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:21.283 "dma_device_type": 2 00:11:21.283 } 00:11:21.283 ], 00:11:21.283 "driver_specific": {} 00:11:21.283 } 00:11:21.283 ] 00:11:21.283 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:21.283 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:11:21.283 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:11:21.283 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:21.283 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:11:21.283 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:21.283 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.283 BaseBdev3 00:11:21.283 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:21.283 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:11:21.283 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:11:21.283 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:21.283 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:11:21.283 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:21.283 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:21.283 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:21.283 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:21.283 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.283 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:21.283 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:11:21.283 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:21.283 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.283 [ 00:11:21.283 { 00:11:21.283 "name": "BaseBdev3", 00:11:21.283 "aliases": [ 00:11:21.283 "05698a71-fa82-4211-8f5f-eacac9e8f8bc" 00:11:21.283 ], 00:11:21.283 "product_name": "Malloc disk", 00:11:21.283 "block_size": 512, 00:11:21.283 "num_blocks": 65536, 00:11:21.283 "uuid": "05698a71-fa82-4211-8f5f-eacac9e8f8bc", 00:11:21.283 "assigned_rate_limits": { 00:11:21.283 "rw_ios_per_sec": 0, 00:11:21.283 "rw_mbytes_per_sec": 0, 00:11:21.283 "r_mbytes_per_sec": 0, 00:11:21.283 "w_mbytes_per_sec": 0 00:11:21.283 }, 00:11:21.283 "claimed": false, 00:11:21.283 "zoned": false, 00:11:21.283 "supported_io_types": { 00:11:21.283 "read": true, 00:11:21.283 "write": true, 00:11:21.283 "unmap": true, 00:11:21.283 "flush": true, 00:11:21.283 "reset": true, 00:11:21.283 "nvme_admin": false, 00:11:21.283 "nvme_io": false, 00:11:21.283 "nvme_io_md": false, 00:11:21.283 "write_zeroes": true, 00:11:21.283 "zcopy": true, 00:11:21.283 "get_zone_info": false, 00:11:21.283 "zone_management": false, 00:11:21.283 "zone_append": false, 00:11:21.283 "compare": false, 00:11:21.283 "compare_and_write": false, 00:11:21.283 "abort": true, 00:11:21.283 "seek_hole": false, 00:11:21.283 "seek_data": false, 00:11:21.283 "copy": true, 00:11:21.283 "nvme_iov_md": false 00:11:21.283 }, 00:11:21.283 "memory_domains": [ 00:11:21.283 { 00:11:21.283 "dma_device_id": "system", 00:11:21.283 "dma_device_type": 1 00:11:21.283 }, 00:11:21.283 { 00:11:21.283 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:21.283 "dma_device_type": 2 00:11:21.283 } 00:11:21.283 ], 00:11:21.283 "driver_specific": {} 00:11:21.283 } 00:11:21.283 ] 00:11:21.283 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:21.283 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:11:21.283 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:11:21.283 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:21.283 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:11:21.283 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:21.283 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.283 BaseBdev4 00:11:21.283 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:21.283 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:11:21.283 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:11:21.283 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:21.283 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:11:21.283 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:21.283 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:21.283 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:21.283 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:21.284 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.284 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:21.284 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:11:21.284 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:21.284 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.284 [ 00:11:21.284 { 00:11:21.284 "name": "BaseBdev4", 00:11:21.284 "aliases": [ 00:11:21.284 "29c73fd7-1351-4c39-a36d-89757a5c1f55" 00:11:21.284 ], 00:11:21.284 "product_name": "Malloc disk", 00:11:21.284 "block_size": 512, 00:11:21.284 "num_blocks": 65536, 00:11:21.284 "uuid": "29c73fd7-1351-4c39-a36d-89757a5c1f55", 00:11:21.284 "assigned_rate_limits": { 00:11:21.284 "rw_ios_per_sec": 0, 00:11:21.284 "rw_mbytes_per_sec": 0, 00:11:21.284 "r_mbytes_per_sec": 0, 00:11:21.284 "w_mbytes_per_sec": 0 00:11:21.284 }, 00:11:21.284 "claimed": false, 00:11:21.284 "zoned": false, 00:11:21.284 "supported_io_types": { 00:11:21.284 "read": true, 00:11:21.284 "write": true, 00:11:21.284 "unmap": true, 00:11:21.284 "flush": true, 00:11:21.284 "reset": true, 00:11:21.284 "nvme_admin": false, 00:11:21.284 "nvme_io": false, 00:11:21.284 "nvme_io_md": false, 00:11:21.284 "write_zeroes": true, 00:11:21.284 "zcopy": true, 00:11:21.284 "get_zone_info": false, 00:11:21.284 "zone_management": false, 00:11:21.284 "zone_append": false, 00:11:21.284 "compare": false, 00:11:21.284 "compare_and_write": false, 00:11:21.284 "abort": true, 00:11:21.284 "seek_hole": false, 00:11:21.284 "seek_data": false, 00:11:21.284 "copy": true, 00:11:21.284 "nvme_iov_md": false 00:11:21.284 }, 00:11:21.284 "memory_domains": [ 00:11:21.284 { 00:11:21.284 "dma_device_id": "system", 00:11:21.284 "dma_device_type": 1 00:11:21.284 }, 00:11:21.284 { 00:11:21.284 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:21.284 "dma_device_type": 2 00:11:21.284 } 00:11:21.284 ], 00:11:21.284 "driver_specific": {} 00:11:21.284 } 00:11:21.284 ] 00:11:21.284 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:21.284 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:11:21.284 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:11:21.284 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:21.284 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:21.284 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:21.284 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.284 [2024-11-19 15:17:11.498560] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:11:21.284 [2024-11-19 15:17:11.498687] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:11:21.284 [2024-11-19 15:17:11.498750] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:21.284 [2024-11-19 15:17:11.500847] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:21.284 [2024-11-19 15:17:11.500938] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:21.284 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:21.284 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:21.284 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:21.284 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:21.284 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:21.284 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:21.284 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:21.284 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:21.284 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:21.284 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:21.284 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:21.284 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:21.284 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:21.284 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.284 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:21.284 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:21.284 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:21.284 "name": "Existed_Raid", 00:11:21.284 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:21.284 "strip_size_kb": 64, 00:11:21.284 "state": "configuring", 00:11:21.284 "raid_level": "concat", 00:11:21.284 "superblock": false, 00:11:21.284 "num_base_bdevs": 4, 00:11:21.284 "num_base_bdevs_discovered": 3, 00:11:21.284 "num_base_bdevs_operational": 4, 00:11:21.284 "base_bdevs_list": [ 00:11:21.284 { 00:11:21.284 "name": "BaseBdev1", 00:11:21.284 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:21.284 "is_configured": false, 00:11:21.284 "data_offset": 0, 00:11:21.284 "data_size": 0 00:11:21.284 }, 00:11:21.284 { 00:11:21.284 "name": "BaseBdev2", 00:11:21.284 "uuid": "858748e1-8394-4f4e-a462-ecf04a6dd41c", 00:11:21.284 "is_configured": true, 00:11:21.284 "data_offset": 0, 00:11:21.284 "data_size": 65536 00:11:21.284 }, 00:11:21.284 { 00:11:21.284 "name": "BaseBdev3", 00:11:21.284 "uuid": "05698a71-fa82-4211-8f5f-eacac9e8f8bc", 00:11:21.284 "is_configured": true, 00:11:21.284 "data_offset": 0, 00:11:21.284 "data_size": 65536 00:11:21.284 }, 00:11:21.284 { 00:11:21.284 "name": "BaseBdev4", 00:11:21.284 "uuid": "29c73fd7-1351-4c39-a36d-89757a5c1f55", 00:11:21.284 "is_configured": true, 00:11:21.284 "data_offset": 0, 00:11:21.284 "data_size": 65536 00:11:21.284 } 00:11:21.284 ] 00:11:21.284 }' 00:11:21.284 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:21.284 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.853 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:11:21.853 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:21.853 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.853 [2024-11-19 15:17:11.973766] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:11:21.853 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:21.853 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:21.853 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:21.853 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:21.853 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:21.853 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:21.853 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:21.853 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:21.853 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:21.853 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:21.853 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:21.853 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:21.853 15:17:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:21.853 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:21.853 15:17:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.853 15:17:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:21.853 15:17:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:21.853 "name": "Existed_Raid", 00:11:21.853 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:21.853 "strip_size_kb": 64, 00:11:21.853 "state": "configuring", 00:11:21.853 "raid_level": "concat", 00:11:21.853 "superblock": false, 00:11:21.853 "num_base_bdevs": 4, 00:11:21.853 "num_base_bdevs_discovered": 2, 00:11:21.853 "num_base_bdevs_operational": 4, 00:11:21.853 "base_bdevs_list": [ 00:11:21.853 { 00:11:21.853 "name": "BaseBdev1", 00:11:21.853 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:21.853 "is_configured": false, 00:11:21.853 "data_offset": 0, 00:11:21.853 "data_size": 0 00:11:21.853 }, 00:11:21.853 { 00:11:21.853 "name": null, 00:11:21.853 "uuid": "858748e1-8394-4f4e-a462-ecf04a6dd41c", 00:11:21.853 "is_configured": false, 00:11:21.853 "data_offset": 0, 00:11:21.853 "data_size": 65536 00:11:21.853 }, 00:11:21.853 { 00:11:21.853 "name": "BaseBdev3", 00:11:21.853 "uuid": "05698a71-fa82-4211-8f5f-eacac9e8f8bc", 00:11:21.853 "is_configured": true, 00:11:21.853 "data_offset": 0, 00:11:21.853 "data_size": 65536 00:11:21.853 }, 00:11:21.853 { 00:11:21.853 "name": "BaseBdev4", 00:11:21.853 "uuid": "29c73fd7-1351-4c39-a36d-89757a5c1f55", 00:11:21.853 "is_configured": true, 00:11:21.853 "data_offset": 0, 00:11:21.853 "data_size": 65536 00:11:21.853 } 00:11:21.853 ] 00:11:21.853 }' 00:11:21.853 15:17:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:21.853 15:17:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:22.113 15:17:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:22.113 15:17:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:22.113 15:17:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:22.113 15:17:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:11:22.113 15:17:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:22.113 15:17:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:11:22.113 15:17:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:11:22.113 15:17:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:22.113 15:17:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:22.113 [2024-11-19 15:17:12.445810] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:22.113 BaseBdev1 00:11:22.113 15:17:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:22.113 15:17:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:11:22.114 15:17:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:11:22.114 15:17:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:22.114 15:17:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:11:22.114 15:17:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:22.114 15:17:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:22.114 15:17:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:22.114 15:17:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:22.114 15:17:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:22.374 15:17:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:22.374 15:17:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:11:22.374 15:17:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:22.374 15:17:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:22.374 [ 00:11:22.374 { 00:11:22.374 "name": "BaseBdev1", 00:11:22.374 "aliases": [ 00:11:22.374 "ca4553eb-ea7d-417b-b4f3-44c5fae81dc6" 00:11:22.374 ], 00:11:22.374 "product_name": "Malloc disk", 00:11:22.374 "block_size": 512, 00:11:22.374 "num_blocks": 65536, 00:11:22.374 "uuid": "ca4553eb-ea7d-417b-b4f3-44c5fae81dc6", 00:11:22.374 "assigned_rate_limits": { 00:11:22.374 "rw_ios_per_sec": 0, 00:11:22.374 "rw_mbytes_per_sec": 0, 00:11:22.374 "r_mbytes_per_sec": 0, 00:11:22.374 "w_mbytes_per_sec": 0 00:11:22.374 }, 00:11:22.374 "claimed": true, 00:11:22.374 "claim_type": "exclusive_write", 00:11:22.374 "zoned": false, 00:11:22.374 "supported_io_types": { 00:11:22.374 "read": true, 00:11:22.374 "write": true, 00:11:22.374 "unmap": true, 00:11:22.374 "flush": true, 00:11:22.374 "reset": true, 00:11:22.374 "nvme_admin": false, 00:11:22.374 "nvme_io": false, 00:11:22.374 "nvme_io_md": false, 00:11:22.374 "write_zeroes": true, 00:11:22.374 "zcopy": true, 00:11:22.374 "get_zone_info": false, 00:11:22.374 "zone_management": false, 00:11:22.374 "zone_append": false, 00:11:22.374 "compare": false, 00:11:22.374 "compare_and_write": false, 00:11:22.374 "abort": true, 00:11:22.374 "seek_hole": false, 00:11:22.374 "seek_data": false, 00:11:22.374 "copy": true, 00:11:22.374 "nvme_iov_md": false 00:11:22.374 }, 00:11:22.374 "memory_domains": [ 00:11:22.374 { 00:11:22.374 "dma_device_id": "system", 00:11:22.374 "dma_device_type": 1 00:11:22.374 }, 00:11:22.374 { 00:11:22.374 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:22.374 "dma_device_type": 2 00:11:22.374 } 00:11:22.374 ], 00:11:22.374 "driver_specific": {} 00:11:22.374 } 00:11:22.374 ] 00:11:22.374 15:17:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:22.374 15:17:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:11:22.374 15:17:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:22.374 15:17:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:22.374 15:17:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:22.374 15:17:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:22.374 15:17:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:22.374 15:17:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:22.374 15:17:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:22.374 15:17:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:22.374 15:17:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:22.374 15:17:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:22.374 15:17:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:22.374 15:17:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:22.374 15:17:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:22.374 15:17:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:22.374 15:17:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:22.374 15:17:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:22.374 "name": "Existed_Raid", 00:11:22.374 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:22.374 "strip_size_kb": 64, 00:11:22.374 "state": "configuring", 00:11:22.374 "raid_level": "concat", 00:11:22.374 "superblock": false, 00:11:22.374 "num_base_bdevs": 4, 00:11:22.374 "num_base_bdevs_discovered": 3, 00:11:22.374 "num_base_bdevs_operational": 4, 00:11:22.374 "base_bdevs_list": [ 00:11:22.374 { 00:11:22.374 "name": "BaseBdev1", 00:11:22.374 "uuid": "ca4553eb-ea7d-417b-b4f3-44c5fae81dc6", 00:11:22.374 "is_configured": true, 00:11:22.374 "data_offset": 0, 00:11:22.374 "data_size": 65536 00:11:22.374 }, 00:11:22.374 { 00:11:22.374 "name": null, 00:11:22.374 "uuid": "858748e1-8394-4f4e-a462-ecf04a6dd41c", 00:11:22.374 "is_configured": false, 00:11:22.374 "data_offset": 0, 00:11:22.374 "data_size": 65536 00:11:22.374 }, 00:11:22.374 { 00:11:22.374 "name": "BaseBdev3", 00:11:22.374 "uuid": "05698a71-fa82-4211-8f5f-eacac9e8f8bc", 00:11:22.374 "is_configured": true, 00:11:22.374 "data_offset": 0, 00:11:22.374 "data_size": 65536 00:11:22.374 }, 00:11:22.374 { 00:11:22.374 "name": "BaseBdev4", 00:11:22.374 "uuid": "29c73fd7-1351-4c39-a36d-89757a5c1f55", 00:11:22.374 "is_configured": true, 00:11:22.374 "data_offset": 0, 00:11:22.374 "data_size": 65536 00:11:22.374 } 00:11:22.374 ] 00:11:22.374 }' 00:11:22.374 15:17:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:22.374 15:17:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:22.633 15:17:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:11:22.633 15:17:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:22.633 15:17:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:22.633 15:17:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:22.633 15:17:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:22.893 15:17:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:11:22.893 15:17:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:11:22.893 15:17:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:22.893 15:17:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:22.893 [2024-11-19 15:17:12.984984] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:11:22.893 15:17:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:22.893 15:17:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:22.893 15:17:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:22.893 15:17:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:22.893 15:17:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:22.893 15:17:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:22.893 15:17:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:22.893 15:17:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:22.893 15:17:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:22.893 15:17:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:22.893 15:17:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:22.893 15:17:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:22.893 15:17:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:22.893 15:17:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:22.893 15:17:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:22.893 15:17:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:22.893 15:17:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:22.893 "name": "Existed_Raid", 00:11:22.893 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:22.893 "strip_size_kb": 64, 00:11:22.893 "state": "configuring", 00:11:22.893 "raid_level": "concat", 00:11:22.893 "superblock": false, 00:11:22.893 "num_base_bdevs": 4, 00:11:22.893 "num_base_bdevs_discovered": 2, 00:11:22.893 "num_base_bdevs_operational": 4, 00:11:22.893 "base_bdevs_list": [ 00:11:22.893 { 00:11:22.893 "name": "BaseBdev1", 00:11:22.893 "uuid": "ca4553eb-ea7d-417b-b4f3-44c5fae81dc6", 00:11:22.893 "is_configured": true, 00:11:22.893 "data_offset": 0, 00:11:22.893 "data_size": 65536 00:11:22.893 }, 00:11:22.893 { 00:11:22.893 "name": null, 00:11:22.893 "uuid": "858748e1-8394-4f4e-a462-ecf04a6dd41c", 00:11:22.893 "is_configured": false, 00:11:22.893 "data_offset": 0, 00:11:22.893 "data_size": 65536 00:11:22.893 }, 00:11:22.893 { 00:11:22.893 "name": null, 00:11:22.893 "uuid": "05698a71-fa82-4211-8f5f-eacac9e8f8bc", 00:11:22.893 "is_configured": false, 00:11:22.893 "data_offset": 0, 00:11:22.893 "data_size": 65536 00:11:22.893 }, 00:11:22.893 { 00:11:22.893 "name": "BaseBdev4", 00:11:22.893 "uuid": "29c73fd7-1351-4c39-a36d-89757a5c1f55", 00:11:22.893 "is_configured": true, 00:11:22.893 "data_offset": 0, 00:11:22.893 "data_size": 65536 00:11:22.893 } 00:11:22.893 ] 00:11:22.893 }' 00:11:22.893 15:17:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:22.893 15:17:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.153 15:17:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:23.153 15:17:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:23.153 15:17:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.153 15:17:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:11:23.153 15:17:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:23.153 15:17:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:11:23.153 15:17:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:11:23.153 15:17:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:23.153 15:17:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.153 [2024-11-19 15:17:13.464131] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:23.153 15:17:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:23.153 15:17:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:23.153 15:17:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:23.153 15:17:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:23.153 15:17:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:23.153 15:17:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:23.153 15:17:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:23.153 15:17:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:23.153 15:17:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:23.153 15:17:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:23.153 15:17:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:23.153 15:17:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:23.153 15:17:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:23.153 15:17:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:23.154 15:17:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.413 15:17:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:23.413 15:17:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:23.413 "name": "Existed_Raid", 00:11:23.413 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:23.413 "strip_size_kb": 64, 00:11:23.413 "state": "configuring", 00:11:23.413 "raid_level": "concat", 00:11:23.413 "superblock": false, 00:11:23.413 "num_base_bdevs": 4, 00:11:23.413 "num_base_bdevs_discovered": 3, 00:11:23.413 "num_base_bdevs_operational": 4, 00:11:23.413 "base_bdevs_list": [ 00:11:23.413 { 00:11:23.413 "name": "BaseBdev1", 00:11:23.413 "uuid": "ca4553eb-ea7d-417b-b4f3-44c5fae81dc6", 00:11:23.413 "is_configured": true, 00:11:23.413 "data_offset": 0, 00:11:23.413 "data_size": 65536 00:11:23.413 }, 00:11:23.413 { 00:11:23.413 "name": null, 00:11:23.413 "uuid": "858748e1-8394-4f4e-a462-ecf04a6dd41c", 00:11:23.413 "is_configured": false, 00:11:23.413 "data_offset": 0, 00:11:23.413 "data_size": 65536 00:11:23.413 }, 00:11:23.413 { 00:11:23.413 "name": "BaseBdev3", 00:11:23.413 "uuid": "05698a71-fa82-4211-8f5f-eacac9e8f8bc", 00:11:23.413 "is_configured": true, 00:11:23.413 "data_offset": 0, 00:11:23.413 "data_size": 65536 00:11:23.413 }, 00:11:23.413 { 00:11:23.413 "name": "BaseBdev4", 00:11:23.413 "uuid": "29c73fd7-1351-4c39-a36d-89757a5c1f55", 00:11:23.413 "is_configured": true, 00:11:23.413 "data_offset": 0, 00:11:23.413 "data_size": 65536 00:11:23.413 } 00:11:23.413 ] 00:11:23.413 }' 00:11:23.413 15:17:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:23.413 15:17:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.673 15:17:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:23.673 15:17:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:23.673 15:17:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.673 15:17:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:11:23.673 15:17:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:23.673 15:17:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:11:23.673 15:17:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:11:23.673 15:17:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:23.673 15:17:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.673 [2024-11-19 15:17:13.903500] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:23.673 15:17:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:23.673 15:17:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:23.673 15:17:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:23.673 15:17:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:23.673 15:17:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:23.673 15:17:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:23.673 15:17:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:23.673 15:17:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:23.673 15:17:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:23.673 15:17:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:23.673 15:17:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:23.673 15:17:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:23.673 15:17:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:23.673 15:17:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.673 15:17:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:23.673 15:17:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:23.673 15:17:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:23.673 "name": "Existed_Raid", 00:11:23.673 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:23.673 "strip_size_kb": 64, 00:11:23.673 "state": "configuring", 00:11:23.673 "raid_level": "concat", 00:11:23.673 "superblock": false, 00:11:23.673 "num_base_bdevs": 4, 00:11:23.673 "num_base_bdevs_discovered": 2, 00:11:23.673 "num_base_bdevs_operational": 4, 00:11:23.673 "base_bdevs_list": [ 00:11:23.673 { 00:11:23.673 "name": null, 00:11:23.673 "uuid": "ca4553eb-ea7d-417b-b4f3-44c5fae81dc6", 00:11:23.673 "is_configured": false, 00:11:23.673 "data_offset": 0, 00:11:23.673 "data_size": 65536 00:11:23.673 }, 00:11:23.673 { 00:11:23.673 "name": null, 00:11:23.673 "uuid": "858748e1-8394-4f4e-a462-ecf04a6dd41c", 00:11:23.673 "is_configured": false, 00:11:23.673 "data_offset": 0, 00:11:23.673 "data_size": 65536 00:11:23.673 }, 00:11:23.673 { 00:11:23.673 "name": "BaseBdev3", 00:11:23.673 "uuid": "05698a71-fa82-4211-8f5f-eacac9e8f8bc", 00:11:23.673 "is_configured": true, 00:11:23.673 "data_offset": 0, 00:11:23.673 "data_size": 65536 00:11:23.673 }, 00:11:23.673 { 00:11:23.673 "name": "BaseBdev4", 00:11:23.673 "uuid": "29c73fd7-1351-4c39-a36d-89757a5c1f55", 00:11:23.673 "is_configured": true, 00:11:23.673 "data_offset": 0, 00:11:23.673 "data_size": 65536 00:11:23.673 } 00:11:23.673 ] 00:11:23.673 }' 00:11:23.673 15:17:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:23.673 15:17:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.243 15:17:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:24.243 15:17:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:24.243 15:17:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:11:24.243 15:17:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.243 15:17:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:24.243 15:17:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:11:24.243 15:17:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:11:24.243 15:17:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:24.243 15:17:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.243 [2024-11-19 15:17:14.398412] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:24.243 15:17:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:24.243 15:17:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:24.243 15:17:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:24.243 15:17:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:24.243 15:17:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:24.243 15:17:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:24.243 15:17:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:24.243 15:17:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:24.243 15:17:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:24.243 15:17:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:24.243 15:17:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:24.243 15:17:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:24.243 15:17:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:24.243 15:17:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.243 15:17:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:24.243 15:17:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:24.243 15:17:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:24.243 "name": "Existed_Raid", 00:11:24.243 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:24.243 "strip_size_kb": 64, 00:11:24.243 "state": "configuring", 00:11:24.243 "raid_level": "concat", 00:11:24.243 "superblock": false, 00:11:24.243 "num_base_bdevs": 4, 00:11:24.243 "num_base_bdevs_discovered": 3, 00:11:24.243 "num_base_bdevs_operational": 4, 00:11:24.243 "base_bdevs_list": [ 00:11:24.243 { 00:11:24.243 "name": null, 00:11:24.243 "uuid": "ca4553eb-ea7d-417b-b4f3-44c5fae81dc6", 00:11:24.243 "is_configured": false, 00:11:24.243 "data_offset": 0, 00:11:24.243 "data_size": 65536 00:11:24.243 }, 00:11:24.243 { 00:11:24.243 "name": "BaseBdev2", 00:11:24.243 "uuid": "858748e1-8394-4f4e-a462-ecf04a6dd41c", 00:11:24.243 "is_configured": true, 00:11:24.243 "data_offset": 0, 00:11:24.243 "data_size": 65536 00:11:24.243 }, 00:11:24.243 { 00:11:24.243 "name": "BaseBdev3", 00:11:24.243 "uuid": "05698a71-fa82-4211-8f5f-eacac9e8f8bc", 00:11:24.243 "is_configured": true, 00:11:24.243 "data_offset": 0, 00:11:24.243 "data_size": 65536 00:11:24.243 }, 00:11:24.243 { 00:11:24.243 "name": "BaseBdev4", 00:11:24.243 "uuid": "29c73fd7-1351-4c39-a36d-89757a5c1f55", 00:11:24.243 "is_configured": true, 00:11:24.243 "data_offset": 0, 00:11:24.243 "data_size": 65536 00:11:24.243 } 00:11:24.243 ] 00:11:24.243 }' 00:11:24.243 15:17:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:24.243 15:17:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.503 15:17:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:24.503 15:17:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:24.503 15:17:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.503 15:17:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:11:24.763 15:17:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:24.763 15:17:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:11:24.763 15:17:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:24.763 15:17:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:24.763 15:17:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.763 15:17:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:11:24.763 15:17:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:24.763 15:17:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u ca4553eb-ea7d-417b-b4f3-44c5fae81dc6 00:11:24.763 15:17:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:24.763 15:17:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.763 [2024-11-19 15:17:14.934284] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:11:24.763 [2024-11-19 15:17:14.934355] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:11:24.763 [2024-11-19 15:17:14.934364] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:11:24.763 [2024-11-19 15:17:14.934633] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:11:24.763 [2024-11-19 15:17:14.934753] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:11:24.763 [2024-11-19 15:17:14.934764] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:11:24.763 [2024-11-19 15:17:14.934952] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:24.763 NewBaseBdev 00:11:24.763 15:17:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:24.763 15:17:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:11:24.763 15:17:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:11:24.763 15:17:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:24.763 15:17:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:11:24.763 15:17:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:24.763 15:17:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:24.763 15:17:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:24.763 15:17:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:24.763 15:17:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.763 15:17:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:24.763 15:17:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:11:24.763 15:17:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:24.763 15:17:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.763 [ 00:11:24.763 { 00:11:24.763 "name": "NewBaseBdev", 00:11:24.763 "aliases": [ 00:11:24.763 "ca4553eb-ea7d-417b-b4f3-44c5fae81dc6" 00:11:24.763 ], 00:11:24.763 "product_name": "Malloc disk", 00:11:24.763 "block_size": 512, 00:11:24.763 "num_blocks": 65536, 00:11:24.763 "uuid": "ca4553eb-ea7d-417b-b4f3-44c5fae81dc6", 00:11:24.763 "assigned_rate_limits": { 00:11:24.763 "rw_ios_per_sec": 0, 00:11:24.763 "rw_mbytes_per_sec": 0, 00:11:24.763 "r_mbytes_per_sec": 0, 00:11:24.763 "w_mbytes_per_sec": 0 00:11:24.764 }, 00:11:24.764 "claimed": true, 00:11:24.764 "claim_type": "exclusive_write", 00:11:24.764 "zoned": false, 00:11:24.764 "supported_io_types": { 00:11:24.764 "read": true, 00:11:24.764 "write": true, 00:11:24.764 "unmap": true, 00:11:24.764 "flush": true, 00:11:24.764 "reset": true, 00:11:24.764 "nvme_admin": false, 00:11:24.764 "nvme_io": false, 00:11:24.764 "nvme_io_md": false, 00:11:24.764 "write_zeroes": true, 00:11:24.764 "zcopy": true, 00:11:24.764 "get_zone_info": false, 00:11:24.764 "zone_management": false, 00:11:24.764 "zone_append": false, 00:11:24.764 "compare": false, 00:11:24.764 "compare_and_write": false, 00:11:24.764 "abort": true, 00:11:24.764 "seek_hole": false, 00:11:24.764 "seek_data": false, 00:11:24.764 "copy": true, 00:11:24.764 "nvme_iov_md": false 00:11:24.764 }, 00:11:24.764 "memory_domains": [ 00:11:24.764 { 00:11:24.764 "dma_device_id": "system", 00:11:24.764 "dma_device_type": 1 00:11:24.764 }, 00:11:24.764 { 00:11:24.764 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:24.764 "dma_device_type": 2 00:11:24.764 } 00:11:24.764 ], 00:11:24.764 "driver_specific": {} 00:11:24.764 } 00:11:24.764 ] 00:11:24.764 15:17:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:24.764 15:17:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:11:24.764 15:17:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:11:24.764 15:17:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:24.764 15:17:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:24.764 15:17:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:24.764 15:17:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:24.764 15:17:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:24.764 15:17:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:24.764 15:17:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:24.764 15:17:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:24.764 15:17:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:24.764 15:17:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:24.764 15:17:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:24.764 15:17:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:24.764 15:17:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.764 15:17:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:24.764 15:17:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:24.764 "name": "Existed_Raid", 00:11:24.764 "uuid": "de75018f-d069-4659-a48d-a192ded29dd7", 00:11:24.764 "strip_size_kb": 64, 00:11:24.764 "state": "online", 00:11:24.764 "raid_level": "concat", 00:11:24.764 "superblock": false, 00:11:24.764 "num_base_bdevs": 4, 00:11:24.764 "num_base_bdevs_discovered": 4, 00:11:24.764 "num_base_bdevs_operational": 4, 00:11:24.764 "base_bdevs_list": [ 00:11:24.764 { 00:11:24.764 "name": "NewBaseBdev", 00:11:24.764 "uuid": "ca4553eb-ea7d-417b-b4f3-44c5fae81dc6", 00:11:24.764 "is_configured": true, 00:11:24.764 "data_offset": 0, 00:11:24.764 "data_size": 65536 00:11:24.764 }, 00:11:24.764 { 00:11:24.764 "name": "BaseBdev2", 00:11:24.764 "uuid": "858748e1-8394-4f4e-a462-ecf04a6dd41c", 00:11:24.764 "is_configured": true, 00:11:24.764 "data_offset": 0, 00:11:24.764 "data_size": 65536 00:11:24.764 }, 00:11:24.764 { 00:11:24.764 "name": "BaseBdev3", 00:11:24.764 "uuid": "05698a71-fa82-4211-8f5f-eacac9e8f8bc", 00:11:24.764 "is_configured": true, 00:11:24.764 "data_offset": 0, 00:11:24.764 "data_size": 65536 00:11:24.764 }, 00:11:24.764 { 00:11:24.764 "name": "BaseBdev4", 00:11:24.764 "uuid": "29c73fd7-1351-4c39-a36d-89757a5c1f55", 00:11:24.764 "is_configured": true, 00:11:24.764 "data_offset": 0, 00:11:24.764 "data_size": 65536 00:11:24.764 } 00:11:24.764 ] 00:11:24.764 }' 00:11:24.764 15:17:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:24.764 15:17:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.333 15:17:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:11:25.333 15:17:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:11:25.333 15:17:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:11:25.333 15:17:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:11:25.333 15:17:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:11:25.333 15:17:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:11:25.333 15:17:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:11:25.333 15:17:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:25.333 15:17:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.333 15:17:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:11:25.333 [2024-11-19 15:17:15.433844] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:25.333 15:17:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:25.333 15:17:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:11:25.333 "name": "Existed_Raid", 00:11:25.333 "aliases": [ 00:11:25.333 "de75018f-d069-4659-a48d-a192ded29dd7" 00:11:25.333 ], 00:11:25.333 "product_name": "Raid Volume", 00:11:25.333 "block_size": 512, 00:11:25.333 "num_blocks": 262144, 00:11:25.333 "uuid": "de75018f-d069-4659-a48d-a192ded29dd7", 00:11:25.333 "assigned_rate_limits": { 00:11:25.333 "rw_ios_per_sec": 0, 00:11:25.333 "rw_mbytes_per_sec": 0, 00:11:25.333 "r_mbytes_per_sec": 0, 00:11:25.333 "w_mbytes_per_sec": 0 00:11:25.333 }, 00:11:25.333 "claimed": false, 00:11:25.333 "zoned": false, 00:11:25.333 "supported_io_types": { 00:11:25.333 "read": true, 00:11:25.333 "write": true, 00:11:25.333 "unmap": true, 00:11:25.333 "flush": true, 00:11:25.333 "reset": true, 00:11:25.333 "nvme_admin": false, 00:11:25.333 "nvme_io": false, 00:11:25.333 "nvme_io_md": false, 00:11:25.333 "write_zeroes": true, 00:11:25.333 "zcopy": false, 00:11:25.333 "get_zone_info": false, 00:11:25.333 "zone_management": false, 00:11:25.333 "zone_append": false, 00:11:25.333 "compare": false, 00:11:25.333 "compare_and_write": false, 00:11:25.333 "abort": false, 00:11:25.333 "seek_hole": false, 00:11:25.333 "seek_data": false, 00:11:25.333 "copy": false, 00:11:25.333 "nvme_iov_md": false 00:11:25.333 }, 00:11:25.333 "memory_domains": [ 00:11:25.333 { 00:11:25.333 "dma_device_id": "system", 00:11:25.333 "dma_device_type": 1 00:11:25.333 }, 00:11:25.333 { 00:11:25.333 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:25.333 "dma_device_type": 2 00:11:25.333 }, 00:11:25.333 { 00:11:25.333 "dma_device_id": "system", 00:11:25.333 "dma_device_type": 1 00:11:25.333 }, 00:11:25.333 { 00:11:25.333 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:25.333 "dma_device_type": 2 00:11:25.333 }, 00:11:25.333 { 00:11:25.333 "dma_device_id": "system", 00:11:25.333 "dma_device_type": 1 00:11:25.333 }, 00:11:25.333 { 00:11:25.333 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:25.333 "dma_device_type": 2 00:11:25.333 }, 00:11:25.333 { 00:11:25.333 "dma_device_id": "system", 00:11:25.333 "dma_device_type": 1 00:11:25.333 }, 00:11:25.333 { 00:11:25.333 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:25.333 "dma_device_type": 2 00:11:25.333 } 00:11:25.333 ], 00:11:25.333 "driver_specific": { 00:11:25.333 "raid": { 00:11:25.333 "uuid": "de75018f-d069-4659-a48d-a192ded29dd7", 00:11:25.333 "strip_size_kb": 64, 00:11:25.333 "state": "online", 00:11:25.333 "raid_level": "concat", 00:11:25.333 "superblock": false, 00:11:25.334 "num_base_bdevs": 4, 00:11:25.334 "num_base_bdevs_discovered": 4, 00:11:25.334 "num_base_bdevs_operational": 4, 00:11:25.334 "base_bdevs_list": [ 00:11:25.334 { 00:11:25.334 "name": "NewBaseBdev", 00:11:25.334 "uuid": "ca4553eb-ea7d-417b-b4f3-44c5fae81dc6", 00:11:25.334 "is_configured": true, 00:11:25.334 "data_offset": 0, 00:11:25.334 "data_size": 65536 00:11:25.334 }, 00:11:25.334 { 00:11:25.334 "name": "BaseBdev2", 00:11:25.334 "uuid": "858748e1-8394-4f4e-a462-ecf04a6dd41c", 00:11:25.334 "is_configured": true, 00:11:25.334 "data_offset": 0, 00:11:25.334 "data_size": 65536 00:11:25.334 }, 00:11:25.334 { 00:11:25.334 "name": "BaseBdev3", 00:11:25.334 "uuid": "05698a71-fa82-4211-8f5f-eacac9e8f8bc", 00:11:25.334 "is_configured": true, 00:11:25.334 "data_offset": 0, 00:11:25.334 "data_size": 65536 00:11:25.334 }, 00:11:25.334 { 00:11:25.334 "name": "BaseBdev4", 00:11:25.334 "uuid": "29c73fd7-1351-4c39-a36d-89757a5c1f55", 00:11:25.334 "is_configured": true, 00:11:25.334 "data_offset": 0, 00:11:25.334 "data_size": 65536 00:11:25.334 } 00:11:25.334 ] 00:11:25.334 } 00:11:25.334 } 00:11:25.334 }' 00:11:25.334 15:17:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:11:25.334 15:17:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:11:25.334 BaseBdev2 00:11:25.334 BaseBdev3 00:11:25.334 BaseBdev4' 00:11:25.334 15:17:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:25.334 15:17:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:11:25.334 15:17:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:25.334 15:17:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:11:25.334 15:17:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:25.334 15:17:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:25.334 15:17:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.334 15:17:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:25.334 15:17:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:25.334 15:17:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:25.334 15:17:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:25.334 15:17:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:11:25.334 15:17:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:25.334 15:17:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.334 15:17:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:25.334 15:17:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:25.334 15:17:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:25.334 15:17:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:25.334 15:17:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:25.594 15:17:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:11:25.594 15:17:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:25.594 15:17:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:25.594 15:17:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.594 15:17:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:25.594 15:17:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:25.594 15:17:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:25.594 15:17:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:25.594 15:17:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:25.594 15:17:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:11:25.594 15:17:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:25.594 15:17:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.594 15:17:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:25.594 15:17:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:25.594 15:17:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:25.594 15:17:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:11:25.594 15:17:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:25.594 15:17:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.594 [2024-11-19 15:17:15.772921] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:11:25.594 [2024-11-19 15:17:15.772993] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:25.594 [2024-11-19 15:17:15.773077] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:25.594 [2024-11-19 15:17:15.773155] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:25.594 [2024-11-19 15:17:15.773165] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:11:25.594 15:17:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:25.594 15:17:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 82181 00:11:25.595 15:17:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 82181 ']' 00:11:25.595 15:17:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 82181 00:11:25.595 15:17:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:11:25.595 15:17:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:11:25.595 15:17:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 82181 00:11:25.595 15:17:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:11:25.595 15:17:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:11:25.595 15:17:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 82181' 00:11:25.595 killing process with pid 82181 00:11:25.595 15:17:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 82181 00:11:25.595 [2024-11-19 15:17:15.814008] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:25.595 15:17:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 82181 00:11:25.595 [2024-11-19 15:17:15.889898] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:26.165 15:17:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:11:26.165 00:11:26.165 real 0m9.778s 00:11:26.165 user 0m16.408s 00:11:26.165 sys 0m2.086s 00:11:26.165 15:17:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:26.165 15:17:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:26.165 ************************************ 00:11:26.165 END TEST raid_state_function_test 00:11:26.165 ************************************ 00:11:26.165 15:17:16 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test concat 4 true 00:11:26.165 15:17:16 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:11:26.165 15:17:16 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:26.165 15:17:16 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:26.165 ************************************ 00:11:26.165 START TEST raid_state_function_test_sb 00:11:26.165 ************************************ 00:11:26.165 15:17:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test concat 4 true 00:11:26.165 15:17:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:11:26.165 15:17:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:11:26.165 15:17:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:11:26.165 15:17:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:11:26.165 15:17:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:11:26.165 15:17:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:26.165 15:17:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:11:26.165 15:17:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:11:26.165 15:17:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:26.165 15:17:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:11:26.165 15:17:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:11:26.165 15:17:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:26.165 15:17:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:11:26.165 15:17:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:11:26.165 15:17:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:26.165 15:17:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:11:26.165 15:17:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:11:26.165 15:17:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:26.165 15:17:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:11:26.165 15:17:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:11:26.165 15:17:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:11:26.165 15:17:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:11:26.165 15:17:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:11:26.165 15:17:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:11:26.165 15:17:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:11:26.165 15:17:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:11:26.165 15:17:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:11:26.165 15:17:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:11:26.165 15:17:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:11:26.165 Process raid pid: 82834 00:11:26.165 15:17:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=82834 00:11:26.165 15:17:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:11:26.165 15:17:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 82834' 00:11:26.165 15:17:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 82834 00:11:26.165 15:17:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 82834 ']' 00:11:26.165 15:17:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:26.165 15:17:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:26.165 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:26.165 15:17:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:26.165 15:17:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:26.165 15:17:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:26.165 [2024-11-19 15:17:16.387581] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:11:26.165 [2024-11-19 15:17:16.387737] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:11:26.425 [2024-11-19 15:17:16.543505] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:26.425 [2024-11-19 15:17:16.581140] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:26.425 [2024-11-19 15:17:16.656492] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:26.425 [2024-11-19 15:17:16.656535] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:26.993 15:17:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:26.993 15:17:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:11:26.993 15:17:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:26.993 15:17:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:26.993 15:17:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:26.993 [2024-11-19 15:17:17.223535] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:11:26.993 [2024-11-19 15:17:17.223602] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:11:26.993 [2024-11-19 15:17:17.223612] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:26.993 [2024-11-19 15:17:17.223622] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:26.993 [2024-11-19 15:17:17.223628] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:11:26.993 [2024-11-19 15:17:17.223641] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:11:26.993 [2024-11-19 15:17:17.223646] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:11:26.993 [2024-11-19 15:17:17.223655] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:11:26.993 15:17:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:26.993 15:17:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:26.993 15:17:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:26.993 15:17:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:26.993 15:17:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:26.993 15:17:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:26.993 15:17:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:26.993 15:17:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:26.993 15:17:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:26.993 15:17:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:26.993 15:17:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:26.993 15:17:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:26.994 15:17:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:26.994 15:17:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:26.994 15:17:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:26.994 15:17:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:26.994 15:17:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:26.994 "name": "Existed_Raid", 00:11:26.994 "uuid": "144a22ae-6c20-419e-86a8-3ac9f7ef6df2", 00:11:26.994 "strip_size_kb": 64, 00:11:26.994 "state": "configuring", 00:11:26.994 "raid_level": "concat", 00:11:26.994 "superblock": true, 00:11:26.994 "num_base_bdevs": 4, 00:11:26.994 "num_base_bdevs_discovered": 0, 00:11:26.994 "num_base_bdevs_operational": 4, 00:11:26.994 "base_bdevs_list": [ 00:11:26.994 { 00:11:26.994 "name": "BaseBdev1", 00:11:26.994 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:26.994 "is_configured": false, 00:11:26.994 "data_offset": 0, 00:11:26.994 "data_size": 0 00:11:26.994 }, 00:11:26.994 { 00:11:26.994 "name": "BaseBdev2", 00:11:26.994 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:26.994 "is_configured": false, 00:11:26.994 "data_offset": 0, 00:11:26.994 "data_size": 0 00:11:26.994 }, 00:11:26.994 { 00:11:26.994 "name": "BaseBdev3", 00:11:26.994 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:26.994 "is_configured": false, 00:11:26.994 "data_offset": 0, 00:11:26.994 "data_size": 0 00:11:26.994 }, 00:11:26.994 { 00:11:26.994 "name": "BaseBdev4", 00:11:26.994 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:26.994 "is_configured": false, 00:11:26.994 "data_offset": 0, 00:11:26.994 "data_size": 0 00:11:26.994 } 00:11:26.994 ] 00:11:26.994 }' 00:11:26.994 15:17:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:26.994 15:17:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:27.564 15:17:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:11:27.564 15:17:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:27.564 15:17:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:27.564 [2024-11-19 15:17:17.658639] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:11:27.564 [2024-11-19 15:17:17.658754] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:11:27.564 15:17:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:27.564 15:17:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:27.564 15:17:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:27.564 15:17:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:27.564 [2024-11-19 15:17:17.670658] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:11:27.564 [2024-11-19 15:17:17.670738] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:11:27.564 [2024-11-19 15:17:17.670765] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:27.564 [2024-11-19 15:17:17.670787] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:27.564 [2024-11-19 15:17:17.670804] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:11:27.564 [2024-11-19 15:17:17.670824] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:11:27.564 [2024-11-19 15:17:17.670841] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:11:27.564 [2024-11-19 15:17:17.670860] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:11:27.564 15:17:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:27.564 15:17:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:11:27.564 15:17:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:27.564 15:17:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:27.564 [2024-11-19 15:17:17.697777] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:27.564 BaseBdev1 00:11:27.564 15:17:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:27.564 15:17:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:11:27.564 15:17:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:11:27.564 15:17:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:27.564 15:17:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:11:27.564 15:17:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:27.564 15:17:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:27.564 15:17:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:27.564 15:17:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:27.564 15:17:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:27.564 15:17:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:27.564 15:17:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:11:27.564 15:17:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:27.564 15:17:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:27.564 [ 00:11:27.564 { 00:11:27.564 "name": "BaseBdev1", 00:11:27.564 "aliases": [ 00:11:27.564 "ea0aecf4-9924-40c9-95fb-5fcbb2d99f17" 00:11:27.564 ], 00:11:27.564 "product_name": "Malloc disk", 00:11:27.564 "block_size": 512, 00:11:27.564 "num_blocks": 65536, 00:11:27.564 "uuid": "ea0aecf4-9924-40c9-95fb-5fcbb2d99f17", 00:11:27.564 "assigned_rate_limits": { 00:11:27.564 "rw_ios_per_sec": 0, 00:11:27.564 "rw_mbytes_per_sec": 0, 00:11:27.564 "r_mbytes_per_sec": 0, 00:11:27.564 "w_mbytes_per_sec": 0 00:11:27.564 }, 00:11:27.564 "claimed": true, 00:11:27.564 "claim_type": "exclusive_write", 00:11:27.564 "zoned": false, 00:11:27.564 "supported_io_types": { 00:11:27.564 "read": true, 00:11:27.564 "write": true, 00:11:27.564 "unmap": true, 00:11:27.564 "flush": true, 00:11:27.564 "reset": true, 00:11:27.564 "nvme_admin": false, 00:11:27.564 "nvme_io": false, 00:11:27.564 "nvme_io_md": false, 00:11:27.564 "write_zeroes": true, 00:11:27.564 "zcopy": true, 00:11:27.564 "get_zone_info": false, 00:11:27.564 "zone_management": false, 00:11:27.564 "zone_append": false, 00:11:27.564 "compare": false, 00:11:27.564 "compare_and_write": false, 00:11:27.564 "abort": true, 00:11:27.564 "seek_hole": false, 00:11:27.564 "seek_data": false, 00:11:27.564 "copy": true, 00:11:27.564 "nvme_iov_md": false 00:11:27.564 }, 00:11:27.564 "memory_domains": [ 00:11:27.564 { 00:11:27.564 "dma_device_id": "system", 00:11:27.564 "dma_device_type": 1 00:11:27.564 }, 00:11:27.565 { 00:11:27.565 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:27.565 "dma_device_type": 2 00:11:27.565 } 00:11:27.565 ], 00:11:27.565 "driver_specific": {} 00:11:27.565 } 00:11:27.565 ] 00:11:27.565 15:17:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:27.565 15:17:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:11:27.565 15:17:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:27.565 15:17:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:27.565 15:17:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:27.565 15:17:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:27.565 15:17:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:27.565 15:17:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:27.565 15:17:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:27.565 15:17:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:27.565 15:17:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:27.565 15:17:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:27.565 15:17:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:27.565 15:17:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:27.565 15:17:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:27.565 15:17:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:27.565 15:17:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:27.565 15:17:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:27.565 "name": "Existed_Raid", 00:11:27.565 "uuid": "5ccd9b98-4486-4cc6-8b31-7a65c9f5649f", 00:11:27.565 "strip_size_kb": 64, 00:11:27.565 "state": "configuring", 00:11:27.565 "raid_level": "concat", 00:11:27.565 "superblock": true, 00:11:27.565 "num_base_bdevs": 4, 00:11:27.565 "num_base_bdevs_discovered": 1, 00:11:27.565 "num_base_bdevs_operational": 4, 00:11:27.565 "base_bdevs_list": [ 00:11:27.565 { 00:11:27.565 "name": "BaseBdev1", 00:11:27.565 "uuid": "ea0aecf4-9924-40c9-95fb-5fcbb2d99f17", 00:11:27.565 "is_configured": true, 00:11:27.565 "data_offset": 2048, 00:11:27.565 "data_size": 63488 00:11:27.565 }, 00:11:27.565 { 00:11:27.565 "name": "BaseBdev2", 00:11:27.565 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:27.565 "is_configured": false, 00:11:27.565 "data_offset": 0, 00:11:27.565 "data_size": 0 00:11:27.565 }, 00:11:27.565 { 00:11:27.565 "name": "BaseBdev3", 00:11:27.565 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:27.565 "is_configured": false, 00:11:27.565 "data_offset": 0, 00:11:27.565 "data_size": 0 00:11:27.565 }, 00:11:27.565 { 00:11:27.565 "name": "BaseBdev4", 00:11:27.565 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:27.565 "is_configured": false, 00:11:27.565 "data_offset": 0, 00:11:27.565 "data_size": 0 00:11:27.565 } 00:11:27.565 ] 00:11:27.565 }' 00:11:27.565 15:17:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:27.565 15:17:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:28.134 15:17:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:11:28.134 15:17:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:28.134 15:17:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:28.134 [2024-11-19 15:17:18.173033] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:11:28.134 [2024-11-19 15:17:18.173096] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:11:28.134 15:17:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:28.134 15:17:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:28.134 15:17:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:28.134 15:17:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:28.134 [2024-11-19 15:17:18.185045] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:28.134 [2024-11-19 15:17:18.187233] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:28.134 [2024-11-19 15:17:18.187310] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:28.134 [2024-11-19 15:17:18.187358] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:11:28.134 [2024-11-19 15:17:18.187393] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:11:28.134 [2024-11-19 15:17:18.187422] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:11:28.135 [2024-11-19 15:17:18.187456] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:11:28.135 15:17:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:28.135 15:17:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:11:28.135 15:17:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:28.135 15:17:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:28.135 15:17:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:28.135 15:17:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:28.135 15:17:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:28.135 15:17:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:28.135 15:17:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:28.135 15:17:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:28.135 15:17:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:28.135 15:17:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:28.135 15:17:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:28.135 15:17:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:28.135 15:17:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:28.135 15:17:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:28.135 15:17:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:28.135 15:17:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:28.135 15:17:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:28.135 "name": "Existed_Raid", 00:11:28.135 "uuid": "f9450633-253c-4adc-b347-430cdd4f4cf5", 00:11:28.135 "strip_size_kb": 64, 00:11:28.135 "state": "configuring", 00:11:28.135 "raid_level": "concat", 00:11:28.135 "superblock": true, 00:11:28.135 "num_base_bdevs": 4, 00:11:28.135 "num_base_bdevs_discovered": 1, 00:11:28.135 "num_base_bdevs_operational": 4, 00:11:28.135 "base_bdevs_list": [ 00:11:28.135 { 00:11:28.135 "name": "BaseBdev1", 00:11:28.135 "uuid": "ea0aecf4-9924-40c9-95fb-5fcbb2d99f17", 00:11:28.135 "is_configured": true, 00:11:28.135 "data_offset": 2048, 00:11:28.135 "data_size": 63488 00:11:28.135 }, 00:11:28.135 { 00:11:28.135 "name": "BaseBdev2", 00:11:28.135 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:28.135 "is_configured": false, 00:11:28.135 "data_offset": 0, 00:11:28.135 "data_size": 0 00:11:28.135 }, 00:11:28.135 { 00:11:28.135 "name": "BaseBdev3", 00:11:28.135 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:28.135 "is_configured": false, 00:11:28.135 "data_offset": 0, 00:11:28.135 "data_size": 0 00:11:28.135 }, 00:11:28.135 { 00:11:28.135 "name": "BaseBdev4", 00:11:28.135 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:28.135 "is_configured": false, 00:11:28.135 "data_offset": 0, 00:11:28.135 "data_size": 0 00:11:28.135 } 00:11:28.135 ] 00:11:28.135 }' 00:11:28.135 15:17:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:28.135 15:17:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:28.395 15:17:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:11:28.395 15:17:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:28.395 15:17:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:28.395 [2024-11-19 15:17:18.569024] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:28.395 BaseBdev2 00:11:28.395 15:17:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:28.395 15:17:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:11:28.395 15:17:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:11:28.395 15:17:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:28.395 15:17:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:11:28.395 15:17:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:28.395 15:17:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:28.395 15:17:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:28.395 15:17:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:28.395 15:17:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:28.395 15:17:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:28.395 15:17:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:11:28.395 15:17:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:28.395 15:17:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:28.395 [ 00:11:28.395 { 00:11:28.395 "name": "BaseBdev2", 00:11:28.395 "aliases": [ 00:11:28.395 "34a105f1-9026-4c6b-a45e-3665fb538bb7" 00:11:28.395 ], 00:11:28.395 "product_name": "Malloc disk", 00:11:28.395 "block_size": 512, 00:11:28.395 "num_blocks": 65536, 00:11:28.395 "uuid": "34a105f1-9026-4c6b-a45e-3665fb538bb7", 00:11:28.395 "assigned_rate_limits": { 00:11:28.395 "rw_ios_per_sec": 0, 00:11:28.395 "rw_mbytes_per_sec": 0, 00:11:28.395 "r_mbytes_per_sec": 0, 00:11:28.395 "w_mbytes_per_sec": 0 00:11:28.395 }, 00:11:28.395 "claimed": true, 00:11:28.395 "claim_type": "exclusive_write", 00:11:28.395 "zoned": false, 00:11:28.395 "supported_io_types": { 00:11:28.395 "read": true, 00:11:28.395 "write": true, 00:11:28.395 "unmap": true, 00:11:28.395 "flush": true, 00:11:28.395 "reset": true, 00:11:28.395 "nvme_admin": false, 00:11:28.395 "nvme_io": false, 00:11:28.395 "nvme_io_md": false, 00:11:28.395 "write_zeroes": true, 00:11:28.395 "zcopy": true, 00:11:28.395 "get_zone_info": false, 00:11:28.395 "zone_management": false, 00:11:28.395 "zone_append": false, 00:11:28.395 "compare": false, 00:11:28.395 "compare_and_write": false, 00:11:28.395 "abort": true, 00:11:28.395 "seek_hole": false, 00:11:28.395 "seek_data": false, 00:11:28.395 "copy": true, 00:11:28.395 "nvme_iov_md": false 00:11:28.395 }, 00:11:28.395 "memory_domains": [ 00:11:28.395 { 00:11:28.395 "dma_device_id": "system", 00:11:28.395 "dma_device_type": 1 00:11:28.395 }, 00:11:28.395 { 00:11:28.395 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:28.395 "dma_device_type": 2 00:11:28.395 } 00:11:28.395 ], 00:11:28.395 "driver_specific": {} 00:11:28.395 } 00:11:28.395 ] 00:11:28.395 15:17:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:28.395 15:17:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:11:28.395 15:17:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:11:28.395 15:17:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:28.395 15:17:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:28.395 15:17:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:28.395 15:17:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:28.395 15:17:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:28.395 15:17:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:28.395 15:17:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:28.395 15:17:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:28.395 15:17:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:28.395 15:17:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:28.395 15:17:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:28.395 15:17:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:28.395 15:17:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:28.395 15:17:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:28.395 15:17:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:28.395 15:17:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:28.395 15:17:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:28.395 "name": "Existed_Raid", 00:11:28.395 "uuid": "f9450633-253c-4adc-b347-430cdd4f4cf5", 00:11:28.395 "strip_size_kb": 64, 00:11:28.395 "state": "configuring", 00:11:28.395 "raid_level": "concat", 00:11:28.395 "superblock": true, 00:11:28.395 "num_base_bdevs": 4, 00:11:28.395 "num_base_bdevs_discovered": 2, 00:11:28.395 "num_base_bdevs_operational": 4, 00:11:28.395 "base_bdevs_list": [ 00:11:28.395 { 00:11:28.395 "name": "BaseBdev1", 00:11:28.395 "uuid": "ea0aecf4-9924-40c9-95fb-5fcbb2d99f17", 00:11:28.395 "is_configured": true, 00:11:28.395 "data_offset": 2048, 00:11:28.395 "data_size": 63488 00:11:28.395 }, 00:11:28.395 { 00:11:28.395 "name": "BaseBdev2", 00:11:28.395 "uuid": "34a105f1-9026-4c6b-a45e-3665fb538bb7", 00:11:28.395 "is_configured": true, 00:11:28.395 "data_offset": 2048, 00:11:28.395 "data_size": 63488 00:11:28.395 }, 00:11:28.395 { 00:11:28.395 "name": "BaseBdev3", 00:11:28.395 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:28.395 "is_configured": false, 00:11:28.395 "data_offset": 0, 00:11:28.395 "data_size": 0 00:11:28.395 }, 00:11:28.395 { 00:11:28.395 "name": "BaseBdev4", 00:11:28.395 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:28.395 "is_configured": false, 00:11:28.395 "data_offset": 0, 00:11:28.395 "data_size": 0 00:11:28.395 } 00:11:28.395 ] 00:11:28.395 }' 00:11:28.395 15:17:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:28.395 15:17:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:28.972 15:17:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:11:28.972 15:17:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:28.972 15:17:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:28.972 [2024-11-19 15:17:19.065155] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:28.972 BaseBdev3 00:11:28.972 15:17:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:28.972 15:17:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:11:28.972 15:17:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:11:28.972 15:17:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:28.972 15:17:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:11:28.972 15:17:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:28.972 15:17:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:28.972 15:17:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:28.972 15:17:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:28.972 15:17:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:28.972 15:17:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:28.972 15:17:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:11:28.972 15:17:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:28.972 15:17:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:28.972 [ 00:11:28.972 { 00:11:28.972 "name": "BaseBdev3", 00:11:28.972 "aliases": [ 00:11:28.972 "69aa7b8d-4701-406a-b96e-ce7a6e3be863" 00:11:28.972 ], 00:11:28.972 "product_name": "Malloc disk", 00:11:28.972 "block_size": 512, 00:11:28.972 "num_blocks": 65536, 00:11:28.972 "uuid": "69aa7b8d-4701-406a-b96e-ce7a6e3be863", 00:11:28.972 "assigned_rate_limits": { 00:11:28.972 "rw_ios_per_sec": 0, 00:11:28.972 "rw_mbytes_per_sec": 0, 00:11:28.972 "r_mbytes_per_sec": 0, 00:11:28.972 "w_mbytes_per_sec": 0 00:11:28.972 }, 00:11:28.972 "claimed": true, 00:11:28.972 "claim_type": "exclusive_write", 00:11:28.972 "zoned": false, 00:11:28.972 "supported_io_types": { 00:11:28.972 "read": true, 00:11:28.972 "write": true, 00:11:28.972 "unmap": true, 00:11:28.972 "flush": true, 00:11:28.972 "reset": true, 00:11:28.972 "nvme_admin": false, 00:11:28.972 "nvme_io": false, 00:11:28.972 "nvme_io_md": false, 00:11:28.972 "write_zeroes": true, 00:11:28.972 "zcopy": true, 00:11:28.972 "get_zone_info": false, 00:11:28.972 "zone_management": false, 00:11:28.972 "zone_append": false, 00:11:28.972 "compare": false, 00:11:28.972 "compare_and_write": false, 00:11:28.972 "abort": true, 00:11:28.972 "seek_hole": false, 00:11:28.972 "seek_data": false, 00:11:28.972 "copy": true, 00:11:28.972 "nvme_iov_md": false 00:11:28.972 }, 00:11:28.972 "memory_domains": [ 00:11:28.972 { 00:11:28.972 "dma_device_id": "system", 00:11:28.972 "dma_device_type": 1 00:11:28.972 }, 00:11:28.972 { 00:11:28.972 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:28.972 "dma_device_type": 2 00:11:28.972 } 00:11:28.972 ], 00:11:28.972 "driver_specific": {} 00:11:28.972 } 00:11:28.972 ] 00:11:28.972 15:17:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:28.972 15:17:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:11:28.972 15:17:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:11:28.972 15:17:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:28.972 15:17:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:28.972 15:17:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:28.972 15:17:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:28.972 15:17:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:28.972 15:17:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:28.972 15:17:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:28.972 15:17:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:28.972 15:17:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:28.972 15:17:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:28.972 15:17:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:28.972 15:17:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:28.972 15:17:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:28.972 15:17:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:28.972 15:17:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:28.972 15:17:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:28.972 15:17:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:28.972 "name": "Existed_Raid", 00:11:28.972 "uuid": "f9450633-253c-4adc-b347-430cdd4f4cf5", 00:11:28.972 "strip_size_kb": 64, 00:11:28.972 "state": "configuring", 00:11:28.972 "raid_level": "concat", 00:11:28.972 "superblock": true, 00:11:28.972 "num_base_bdevs": 4, 00:11:28.972 "num_base_bdevs_discovered": 3, 00:11:28.972 "num_base_bdevs_operational": 4, 00:11:28.972 "base_bdevs_list": [ 00:11:28.972 { 00:11:28.972 "name": "BaseBdev1", 00:11:28.972 "uuid": "ea0aecf4-9924-40c9-95fb-5fcbb2d99f17", 00:11:28.972 "is_configured": true, 00:11:28.972 "data_offset": 2048, 00:11:28.972 "data_size": 63488 00:11:28.972 }, 00:11:28.972 { 00:11:28.972 "name": "BaseBdev2", 00:11:28.972 "uuid": "34a105f1-9026-4c6b-a45e-3665fb538bb7", 00:11:28.972 "is_configured": true, 00:11:28.972 "data_offset": 2048, 00:11:28.973 "data_size": 63488 00:11:28.973 }, 00:11:28.973 { 00:11:28.973 "name": "BaseBdev3", 00:11:28.973 "uuid": "69aa7b8d-4701-406a-b96e-ce7a6e3be863", 00:11:28.973 "is_configured": true, 00:11:28.973 "data_offset": 2048, 00:11:28.973 "data_size": 63488 00:11:28.973 }, 00:11:28.973 { 00:11:28.973 "name": "BaseBdev4", 00:11:28.973 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:28.973 "is_configured": false, 00:11:28.973 "data_offset": 0, 00:11:28.973 "data_size": 0 00:11:28.973 } 00:11:28.973 ] 00:11:28.973 }' 00:11:28.973 15:17:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:28.973 15:17:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:29.263 15:17:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:11:29.263 15:17:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:29.263 15:17:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:29.263 [2024-11-19 15:17:19.581058] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:29.263 [2024-11-19 15:17:19.581389] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:11:29.263 [2024-11-19 15:17:19.581446] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:11:29.263 BaseBdev4 00:11:29.263 [2024-11-19 15:17:19.581827] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:11:29.263 [2024-11-19 15:17:19.581993] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:11:29.263 [2024-11-19 15:17:19.582008] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:11:29.263 [2024-11-19 15:17:19.582128] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:29.263 15:17:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:29.263 15:17:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:11:29.263 15:17:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:11:29.263 15:17:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:29.263 15:17:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:11:29.263 15:17:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:29.263 15:17:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:29.263 15:17:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:29.263 15:17:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:29.263 15:17:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:29.263 15:17:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:29.263 15:17:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:11:29.263 15:17:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:29.263 15:17:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:29.523 [ 00:11:29.523 { 00:11:29.523 "name": "BaseBdev4", 00:11:29.523 "aliases": [ 00:11:29.523 "5015a0bc-0a05-4cb8-982a-ae9cd0cce85c" 00:11:29.523 ], 00:11:29.523 "product_name": "Malloc disk", 00:11:29.523 "block_size": 512, 00:11:29.523 "num_blocks": 65536, 00:11:29.523 "uuid": "5015a0bc-0a05-4cb8-982a-ae9cd0cce85c", 00:11:29.523 "assigned_rate_limits": { 00:11:29.523 "rw_ios_per_sec": 0, 00:11:29.523 "rw_mbytes_per_sec": 0, 00:11:29.523 "r_mbytes_per_sec": 0, 00:11:29.523 "w_mbytes_per_sec": 0 00:11:29.523 }, 00:11:29.523 "claimed": true, 00:11:29.523 "claim_type": "exclusive_write", 00:11:29.523 "zoned": false, 00:11:29.523 "supported_io_types": { 00:11:29.523 "read": true, 00:11:29.523 "write": true, 00:11:29.523 "unmap": true, 00:11:29.523 "flush": true, 00:11:29.523 "reset": true, 00:11:29.523 "nvme_admin": false, 00:11:29.523 "nvme_io": false, 00:11:29.523 "nvme_io_md": false, 00:11:29.523 "write_zeroes": true, 00:11:29.523 "zcopy": true, 00:11:29.523 "get_zone_info": false, 00:11:29.523 "zone_management": false, 00:11:29.523 "zone_append": false, 00:11:29.523 "compare": false, 00:11:29.523 "compare_and_write": false, 00:11:29.523 "abort": true, 00:11:29.523 "seek_hole": false, 00:11:29.523 "seek_data": false, 00:11:29.523 "copy": true, 00:11:29.523 "nvme_iov_md": false 00:11:29.523 }, 00:11:29.523 "memory_domains": [ 00:11:29.523 { 00:11:29.523 "dma_device_id": "system", 00:11:29.523 "dma_device_type": 1 00:11:29.523 }, 00:11:29.523 { 00:11:29.523 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:29.523 "dma_device_type": 2 00:11:29.523 } 00:11:29.523 ], 00:11:29.523 "driver_specific": {} 00:11:29.523 } 00:11:29.523 ] 00:11:29.523 15:17:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:29.523 15:17:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:11:29.523 15:17:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:11:29.523 15:17:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:29.523 15:17:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:11:29.523 15:17:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:29.523 15:17:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:29.523 15:17:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:29.523 15:17:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:29.523 15:17:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:29.523 15:17:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:29.523 15:17:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:29.523 15:17:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:29.523 15:17:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:29.523 15:17:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:29.523 15:17:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:29.523 15:17:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:29.523 15:17:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:29.523 15:17:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:29.523 15:17:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:29.523 "name": "Existed_Raid", 00:11:29.523 "uuid": "f9450633-253c-4adc-b347-430cdd4f4cf5", 00:11:29.523 "strip_size_kb": 64, 00:11:29.523 "state": "online", 00:11:29.523 "raid_level": "concat", 00:11:29.523 "superblock": true, 00:11:29.523 "num_base_bdevs": 4, 00:11:29.523 "num_base_bdevs_discovered": 4, 00:11:29.523 "num_base_bdevs_operational": 4, 00:11:29.523 "base_bdevs_list": [ 00:11:29.523 { 00:11:29.523 "name": "BaseBdev1", 00:11:29.523 "uuid": "ea0aecf4-9924-40c9-95fb-5fcbb2d99f17", 00:11:29.523 "is_configured": true, 00:11:29.523 "data_offset": 2048, 00:11:29.523 "data_size": 63488 00:11:29.523 }, 00:11:29.523 { 00:11:29.523 "name": "BaseBdev2", 00:11:29.523 "uuid": "34a105f1-9026-4c6b-a45e-3665fb538bb7", 00:11:29.523 "is_configured": true, 00:11:29.523 "data_offset": 2048, 00:11:29.523 "data_size": 63488 00:11:29.523 }, 00:11:29.523 { 00:11:29.523 "name": "BaseBdev3", 00:11:29.523 "uuid": "69aa7b8d-4701-406a-b96e-ce7a6e3be863", 00:11:29.523 "is_configured": true, 00:11:29.523 "data_offset": 2048, 00:11:29.523 "data_size": 63488 00:11:29.523 }, 00:11:29.523 { 00:11:29.523 "name": "BaseBdev4", 00:11:29.523 "uuid": "5015a0bc-0a05-4cb8-982a-ae9cd0cce85c", 00:11:29.523 "is_configured": true, 00:11:29.523 "data_offset": 2048, 00:11:29.523 "data_size": 63488 00:11:29.523 } 00:11:29.523 ] 00:11:29.523 }' 00:11:29.523 15:17:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:29.523 15:17:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:29.783 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:11:29.783 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:11:29.783 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:11:29.783 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:11:29.783 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:11:29.783 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:11:29.783 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:11:29.783 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:11:29.783 15:17:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:29.783 15:17:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:29.783 [2024-11-19 15:17:20.092511] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:29.783 15:17:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:30.043 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:11:30.043 "name": "Existed_Raid", 00:11:30.043 "aliases": [ 00:11:30.043 "f9450633-253c-4adc-b347-430cdd4f4cf5" 00:11:30.043 ], 00:11:30.043 "product_name": "Raid Volume", 00:11:30.043 "block_size": 512, 00:11:30.043 "num_blocks": 253952, 00:11:30.043 "uuid": "f9450633-253c-4adc-b347-430cdd4f4cf5", 00:11:30.043 "assigned_rate_limits": { 00:11:30.043 "rw_ios_per_sec": 0, 00:11:30.043 "rw_mbytes_per_sec": 0, 00:11:30.043 "r_mbytes_per_sec": 0, 00:11:30.043 "w_mbytes_per_sec": 0 00:11:30.043 }, 00:11:30.043 "claimed": false, 00:11:30.043 "zoned": false, 00:11:30.043 "supported_io_types": { 00:11:30.043 "read": true, 00:11:30.043 "write": true, 00:11:30.043 "unmap": true, 00:11:30.043 "flush": true, 00:11:30.043 "reset": true, 00:11:30.043 "nvme_admin": false, 00:11:30.043 "nvme_io": false, 00:11:30.043 "nvme_io_md": false, 00:11:30.043 "write_zeroes": true, 00:11:30.043 "zcopy": false, 00:11:30.043 "get_zone_info": false, 00:11:30.043 "zone_management": false, 00:11:30.043 "zone_append": false, 00:11:30.043 "compare": false, 00:11:30.043 "compare_and_write": false, 00:11:30.043 "abort": false, 00:11:30.043 "seek_hole": false, 00:11:30.043 "seek_data": false, 00:11:30.043 "copy": false, 00:11:30.043 "nvme_iov_md": false 00:11:30.043 }, 00:11:30.043 "memory_domains": [ 00:11:30.043 { 00:11:30.043 "dma_device_id": "system", 00:11:30.043 "dma_device_type": 1 00:11:30.043 }, 00:11:30.043 { 00:11:30.043 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:30.043 "dma_device_type": 2 00:11:30.043 }, 00:11:30.043 { 00:11:30.043 "dma_device_id": "system", 00:11:30.043 "dma_device_type": 1 00:11:30.043 }, 00:11:30.043 { 00:11:30.043 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:30.043 "dma_device_type": 2 00:11:30.043 }, 00:11:30.043 { 00:11:30.043 "dma_device_id": "system", 00:11:30.043 "dma_device_type": 1 00:11:30.043 }, 00:11:30.043 { 00:11:30.043 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:30.043 "dma_device_type": 2 00:11:30.043 }, 00:11:30.043 { 00:11:30.043 "dma_device_id": "system", 00:11:30.043 "dma_device_type": 1 00:11:30.043 }, 00:11:30.043 { 00:11:30.043 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:30.043 "dma_device_type": 2 00:11:30.043 } 00:11:30.043 ], 00:11:30.043 "driver_specific": { 00:11:30.043 "raid": { 00:11:30.043 "uuid": "f9450633-253c-4adc-b347-430cdd4f4cf5", 00:11:30.043 "strip_size_kb": 64, 00:11:30.043 "state": "online", 00:11:30.043 "raid_level": "concat", 00:11:30.043 "superblock": true, 00:11:30.043 "num_base_bdevs": 4, 00:11:30.043 "num_base_bdevs_discovered": 4, 00:11:30.043 "num_base_bdevs_operational": 4, 00:11:30.043 "base_bdevs_list": [ 00:11:30.043 { 00:11:30.043 "name": "BaseBdev1", 00:11:30.043 "uuid": "ea0aecf4-9924-40c9-95fb-5fcbb2d99f17", 00:11:30.043 "is_configured": true, 00:11:30.043 "data_offset": 2048, 00:11:30.043 "data_size": 63488 00:11:30.043 }, 00:11:30.043 { 00:11:30.043 "name": "BaseBdev2", 00:11:30.043 "uuid": "34a105f1-9026-4c6b-a45e-3665fb538bb7", 00:11:30.043 "is_configured": true, 00:11:30.043 "data_offset": 2048, 00:11:30.043 "data_size": 63488 00:11:30.043 }, 00:11:30.043 { 00:11:30.043 "name": "BaseBdev3", 00:11:30.043 "uuid": "69aa7b8d-4701-406a-b96e-ce7a6e3be863", 00:11:30.043 "is_configured": true, 00:11:30.043 "data_offset": 2048, 00:11:30.043 "data_size": 63488 00:11:30.043 }, 00:11:30.043 { 00:11:30.043 "name": "BaseBdev4", 00:11:30.043 "uuid": "5015a0bc-0a05-4cb8-982a-ae9cd0cce85c", 00:11:30.043 "is_configured": true, 00:11:30.043 "data_offset": 2048, 00:11:30.043 "data_size": 63488 00:11:30.043 } 00:11:30.043 ] 00:11:30.043 } 00:11:30.043 } 00:11:30.043 }' 00:11:30.043 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:11:30.043 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:11:30.043 BaseBdev2 00:11:30.044 BaseBdev3 00:11:30.044 BaseBdev4' 00:11:30.044 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:30.044 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:11:30.044 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:30.044 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:11:30.044 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:30.044 15:17:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:30.044 15:17:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:30.044 15:17:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:30.044 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:30.044 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:30.044 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:30.044 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:11:30.044 15:17:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:30.044 15:17:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:30.044 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:30.044 15:17:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:30.044 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:30.044 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:30.044 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:30.044 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:11:30.044 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:30.044 15:17:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:30.044 15:17:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:30.044 15:17:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:30.044 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:30.044 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:30.044 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:30.044 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:30.044 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:11:30.044 15:17:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:30.044 15:17:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:30.304 15:17:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:30.304 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:30.304 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:30.304 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:11:30.304 15:17:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:30.304 15:17:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:30.304 [2024-11-19 15:17:20.423749] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:30.304 [2024-11-19 15:17:20.423844] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:30.304 [2024-11-19 15:17:20.423925] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:30.304 15:17:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:30.304 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:11:30.304 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:11:30.304 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:11:30.304 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:11:30.304 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:11:30.304 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 3 00:11:30.304 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:30.304 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:11:30.304 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:30.304 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:30.304 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:11:30.304 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:30.304 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:30.304 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:30.304 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:30.304 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:30.304 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:30.304 15:17:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:30.304 15:17:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:30.304 15:17:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:30.304 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:30.304 "name": "Existed_Raid", 00:11:30.304 "uuid": "f9450633-253c-4adc-b347-430cdd4f4cf5", 00:11:30.304 "strip_size_kb": 64, 00:11:30.304 "state": "offline", 00:11:30.304 "raid_level": "concat", 00:11:30.304 "superblock": true, 00:11:30.304 "num_base_bdevs": 4, 00:11:30.304 "num_base_bdevs_discovered": 3, 00:11:30.304 "num_base_bdevs_operational": 3, 00:11:30.304 "base_bdevs_list": [ 00:11:30.304 { 00:11:30.304 "name": null, 00:11:30.304 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:30.304 "is_configured": false, 00:11:30.304 "data_offset": 0, 00:11:30.304 "data_size": 63488 00:11:30.304 }, 00:11:30.304 { 00:11:30.304 "name": "BaseBdev2", 00:11:30.304 "uuid": "34a105f1-9026-4c6b-a45e-3665fb538bb7", 00:11:30.304 "is_configured": true, 00:11:30.304 "data_offset": 2048, 00:11:30.304 "data_size": 63488 00:11:30.304 }, 00:11:30.304 { 00:11:30.304 "name": "BaseBdev3", 00:11:30.304 "uuid": "69aa7b8d-4701-406a-b96e-ce7a6e3be863", 00:11:30.304 "is_configured": true, 00:11:30.304 "data_offset": 2048, 00:11:30.304 "data_size": 63488 00:11:30.304 }, 00:11:30.304 { 00:11:30.304 "name": "BaseBdev4", 00:11:30.304 "uuid": "5015a0bc-0a05-4cb8-982a-ae9cd0cce85c", 00:11:30.304 "is_configured": true, 00:11:30.304 "data_offset": 2048, 00:11:30.304 "data_size": 63488 00:11:30.304 } 00:11:30.304 ] 00:11:30.304 }' 00:11:30.304 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:30.304 15:17:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:30.874 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:11:30.874 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:11:30.874 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:30.874 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:11:30.874 15:17:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:30.874 15:17:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:30.874 15:17:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:30.874 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:11:30.874 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:11:30.874 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:11:30.874 15:17:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:30.874 15:17:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:30.874 [2024-11-19 15:17:20.964061] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:11:30.874 15:17:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:30.874 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:11:30.874 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:11:30.874 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:30.874 15:17:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:30.874 15:17:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:30.874 15:17:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:11:30.874 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:30.874 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:11:30.874 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:11:30.874 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:11:30.874 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:30.874 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:30.874 [2024-11-19 15:17:21.032900] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:11:30.874 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:30.874 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:11:30.874 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:11:30.874 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:30.874 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:11:30.874 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:30.874 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:30.874 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:30.874 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:11:30.874 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:11:30.874 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:11:30.874 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:30.874 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:30.874 [2024-11-19 15:17:21.113312] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:11:30.874 [2024-11-19 15:17:21.113442] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:11:30.874 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:30.874 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:11:30.874 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:11:30.875 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:30.875 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:11:30.875 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:30.875 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:30.875 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:30.875 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:11:30.875 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:11:30.875 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:11:30.875 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:11:30.875 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:30.875 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:11:30.875 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:30.875 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:31.135 BaseBdev2 00:11:31.135 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:31.135 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:11:31.135 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:11:31.135 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:31.135 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:11:31.135 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:31.135 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:31.135 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:31.135 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:31.135 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:31.135 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:31.135 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:11:31.135 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:31.135 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:31.135 [ 00:11:31.135 { 00:11:31.135 "name": "BaseBdev2", 00:11:31.135 "aliases": [ 00:11:31.135 "cdf7dc43-301f-450a-87ed-2e0c2bf0b35d" 00:11:31.135 ], 00:11:31.135 "product_name": "Malloc disk", 00:11:31.135 "block_size": 512, 00:11:31.135 "num_blocks": 65536, 00:11:31.135 "uuid": "cdf7dc43-301f-450a-87ed-2e0c2bf0b35d", 00:11:31.135 "assigned_rate_limits": { 00:11:31.135 "rw_ios_per_sec": 0, 00:11:31.135 "rw_mbytes_per_sec": 0, 00:11:31.135 "r_mbytes_per_sec": 0, 00:11:31.135 "w_mbytes_per_sec": 0 00:11:31.135 }, 00:11:31.135 "claimed": false, 00:11:31.135 "zoned": false, 00:11:31.135 "supported_io_types": { 00:11:31.135 "read": true, 00:11:31.135 "write": true, 00:11:31.135 "unmap": true, 00:11:31.135 "flush": true, 00:11:31.135 "reset": true, 00:11:31.135 "nvme_admin": false, 00:11:31.135 "nvme_io": false, 00:11:31.135 "nvme_io_md": false, 00:11:31.135 "write_zeroes": true, 00:11:31.135 "zcopy": true, 00:11:31.135 "get_zone_info": false, 00:11:31.135 "zone_management": false, 00:11:31.135 "zone_append": false, 00:11:31.135 "compare": false, 00:11:31.135 "compare_and_write": false, 00:11:31.135 "abort": true, 00:11:31.135 "seek_hole": false, 00:11:31.135 "seek_data": false, 00:11:31.135 "copy": true, 00:11:31.135 "nvme_iov_md": false 00:11:31.135 }, 00:11:31.135 "memory_domains": [ 00:11:31.135 { 00:11:31.135 "dma_device_id": "system", 00:11:31.135 "dma_device_type": 1 00:11:31.135 }, 00:11:31.135 { 00:11:31.135 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:31.135 "dma_device_type": 2 00:11:31.135 } 00:11:31.135 ], 00:11:31.135 "driver_specific": {} 00:11:31.135 } 00:11:31.135 ] 00:11:31.135 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:31.135 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:11:31.135 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:11:31.135 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:31.135 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:11:31.135 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:31.135 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:31.135 BaseBdev3 00:11:31.135 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:31.135 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:31.136 [ 00:11:31.136 { 00:11:31.136 "name": "BaseBdev3", 00:11:31.136 "aliases": [ 00:11:31.136 "0e75a818-d41e-45fd-9944-f503285583aa" 00:11:31.136 ], 00:11:31.136 "product_name": "Malloc disk", 00:11:31.136 "block_size": 512, 00:11:31.136 "num_blocks": 65536, 00:11:31.136 "uuid": "0e75a818-d41e-45fd-9944-f503285583aa", 00:11:31.136 "assigned_rate_limits": { 00:11:31.136 "rw_ios_per_sec": 0, 00:11:31.136 "rw_mbytes_per_sec": 0, 00:11:31.136 "r_mbytes_per_sec": 0, 00:11:31.136 "w_mbytes_per_sec": 0 00:11:31.136 }, 00:11:31.136 "claimed": false, 00:11:31.136 "zoned": false, 00:11:31.136 "supported_io_types": { 00:11:31.136 "read": true, 00:11:31.136 "write": true, 00:11:31.136 "unmap": true, 00:11:31.136 "flush": true, 00:11:31.136 "reset": true, 00:11:31.136 "nvme_admin": false, 00:11:31.136 "nvme_io": false, 00:11:31.136 "nvme_io_md": false, 00:11:31.136 "write_zeroes": true, 00:11:31.136 "zcopy": true, 00:11:31.136 "get_zone_info": false, 00:11:31.136 "zone_management": false, 00:11:31.136 "zone_append": false, 00:11:31.136 "compare": false, 00:11:31.136 "compare_and_write": false, 00:11:31.136 "abort": true, 00:11:31.136 "seek_hole": false, 00:11:31.136 "seek_data": false, 00:11:31.136 "copy": true, 00:11:31.136 "nvme_iov_md": false 00:11:31.136 }, 00:11:31.136 "memory_domains": [ 00:11:31.136 { 00:11:31.136 "dma_device_id": "system", 00:11:31.136 "dma_device_type": 1 00:11:31.136 }, 00:11:31.136 { 00:11:31.136 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:31.136 "dma_device_type": 2 00:11:31.136 } 00:11:31.136 ], 00:11:31.136 "driver_specific": {} 00:11:31.136 } 00:11:31.136 ] 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:31.136 BaseBdev4 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:31.136 [ 00:11:31.136 { 00:11:31.136 "name": "BaseBdev4", 00:11:31.136 "aliases": [ 00:11:31.136 "8510c761-719a-4ea5-a504-27a8025e09ca" 00:11:31.136 ], 00:11:31.136 "product_name": "Malloc disk", 00:11:31.136 "block_size": 512, 00:11:31.136 "num_blocks": 65536, 00:11:31.136 "uuid": "8510c761-719a-4ea5-a504-27a8025e09ca", 00:11:31.136 "assigned_rate_limits": { 00:11:31.136 "rw_ios_per_sec": 0, 00:11:31.136 "rw_mbytes_per_sec": 0, 00:11:31.136 "r_mbytes_per_sec": 0, 00:11:31.136 "w_mbytes_per_sec": 0 00:11:31.136 }, 00:11:31.136 "claimed": false, 00:11:31.136 "zoned": false, 00:11:31.136 "supported_io_types": { 00:11:31.136 "read": true, 00:11:31.136 "write": true, 00:11:31.136 "unmap": true, 00:11:31.136 "flush": true, 00:11:31.136 "reset": true, 00:11:31.136 "nvme_admin": false, 00:11:31.136 "nvme_io": false, 00:11:31.136 "nvme_io_md": false, 00:11:31.136 "write_zeroes": true, 00:11:31.136 "zcopy": true, 00:11:31.136 "get_zone_info": false, 00:11:31.136 "zone_management": false, 00:11:31.136 "zone_append": false, 00:11:31.136 "compare": false, 00:11:31.136 "compare_and_write": false, 00:11:31.136 "abort": true, 00:11:31.136 "seek_hole": false, 00:11:31.136 "seek_data": false, 00:11:31.136 "copy": true, 00:11:31.136 "nvme_iov_md": false 00:11:31.136 }, 00:11:31.136 "memory_domains": [ 00:11:31.136 { 00:11:31.136 "dma_device_id": "system", 00:11:31.136 "dma_device_type": 1 00:11:31.136 }, 00:11:31.136 { 00:11:31.136 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:31.136 "dma_device_type": 2 00:11:31.136 } 00:11:31.136 ], 00:11:31.136 "driver_specific": {} 00:11:31.136 } 00:11:31.136 ] 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:31.136 [2024-11-19 15:17:21.383400] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:11:31.136 [2024-11-19 15:17:21.383534] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:11:31.136 [2024-11-19 15:17:21.383598] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:31.136 [2024-11-19 15:17:21.385781] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:31.136 [2024-11-19 15:17:21.385876] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:31.136 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:31.136 "name": "Existed_Raid", 00:11:31.136 "uuid": "90132b83-6ddd-45e9-b64e-2af0cdb7d01b", 00:11:31.136 "strip_size_kb": 64, 00:11:31.136 "state": "configuring", 00:11:31.136 "raid_level": "concat", 00:11:31.136 "superblock": true, 00:11:31.136 "num_base_bdevs": 4, 00:11:31.136 "num_base_bdevs_discovered": 3, 00:11:31.136 "num_base_bdevs_operational": 4, 00:11:31.136 "base_bdevs_list": [ 00:11:31.136 { 00:11:31.136 "name": "BaseBdev1", 00:11:31.136 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:31.136 "is_configured": false, 00:11:31.136 "data_offset": 0, 00:11:31.137 "data_size": 0 00:11:31.137 }, 00:11:31.137 { 00:11:31.137 "name": "BaseBdev2", 00:11:31.137 "uuid": "cdf7dc43-301f-450a-87ed-2e0c2bf0b35d", 00:11:31.137 "is_configured": true, 00:11:31.137 "data_offset": 2048, 00:11:31.137 "data_size": 63488 00:11:31.137 }, 00:11:31.137 { 00:11:31.137 "name": "BaseBdev3", 00:11:31.137 "uuid": "0e75a818-d41e-45fd-9944-f503285583aa", 00:11:31.137 "is_configured": true, 00:11:31.137 "data_offset": 2048, 00:11:31.137 "data_size": 63488 00:11:31.137 }, 00:11:31.137 { 00:11:31.137 "name": "BaseBdev4", 00:11:31.137 "uuid": "8510c761-719a-4ea5-a504-27a8025e09ca", 00:11:31.137 "is_configured": true, 00:11:31.137 "data_offset": 2048, 00:11:31.137 "data_size": 63488 00:11:31.137 } 00:11:31.137 ] 00:11:31.137 }' 00:11:31.137 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:31.137 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:31.707 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:11:31.707 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:31.707 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:31.707 [2024-11-19 15:17:21.818681] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:11:31.707 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:31.707 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:31.707 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:31.707 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:31.707 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:31.707 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:31.707 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:31.707 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:31.707 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:31.707 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:31.707 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:31.707 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:31.707 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:31.707 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:31.707 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:31.707 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:31.707 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:31.707 "name": "Existed_Raid", 00:11:31.707 "uuid": "90132b83-6ddd-45e9-b64e-2af0cdb7d01b", 00:11:31.707 "strip_size_kb": 64, 00:11:31.707 "state": "configuring", 00:11:31.707 "raid_level": "concat", 00:11:31.707 "superblock": true, 00:11:31.707 "num_base_bdevs": 4, 00:11:31.707 "num_base_bdevs_discovered": 2, 00:11:31.707 "num_base_bdevs_operational": 4, 00:11:31.707 "base_bdevs_list": [ 00:11:31.707 { 00:11:31.707 "name": "BaseBdev1", 00:11:31.707 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:31.707 "is_configured": false, 00:11:31.707 "data_offset": 0, 00:11:31.707 "data_size": 0 00:11:31.707 }, 00:11:31.707 { 00:11:31.707 "name": null, 00:11:31.707 "uuid": "cdf7dc43-301f-450a-87ed-2e0c2bf0b35d", 00:11:31.707 "is_configured": false, 00:11:31.707 "data_offset": 0, 00:11:31.707 "data_size": 63488 00:11:31.707 }, 00:11:31.707 { 00:11:31.707 "name": "BaseBdev3", 00:11:31.707 "uuid": "0e75a818-d41e-45fd-9944-f503285583aa", 00:11:31.707 "is_configured": true, 00:11:31.707 "data_offset": 2048, 00:11:31.707 "data_size": 63488 00:11:31.707 }, 00:11:31.707 { 00:11:31.707 "name": "BaseBdev4", 00:11:31.707 "uuid": "8510c761-719a-4ea5-a504-27a8025e09ca", 00:11:31.707 "is_configured": true, 00:11:31.707 "data_offset": 2048, 00:11:31.707 "data_size": 63488 00:11:31.707 } 00:11:31.707 ] 00:11:31.707 }' 00:11:31.707 15:17:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:31.707 15:17:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:32.278 15:17:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:32.278 15:17:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:32.278 15:17:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:32.278 15:17:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:11:32.278 15:17:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:32.278 15:17:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:11:32.278 15:17:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:11:32.278 15:17:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:32.278 15:17:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:32.278 [2024-11-19 15:17:22.375071] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:32.278 BaseBdev1 00:11:32.278 15:17:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:32.278 15:17:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:11:32.278 15:17:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:11:32.278 15:17:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:32.278 15:17:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:11:32.278 15:17:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:32.278 15:17:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:32.278 15:17:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:32.278 15:17:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:32.278 15:17:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:32.278 15:17:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:32.278 15:17:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:11:32.278 15:17:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:32.278 15:17:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:32.278 [ 00:11:32.278 { 00:11:32.278 "name": "BaseBdev1", 00:11:32.278 "aliases": [ 00:11:32.278 "5b3dd593-6129-4ec0-8d56-4355e82b10d9" 00:11:32.278 ], 00:11:32.278 "product_name": "Malloc disk", 00:11:32.278 "block_size": 512, 00:11:32.278 "num_blocks": 65536, 00:11:32.278 "uuid": "5b3dd593-6129-4ec0-8d56-4355e82b10d9", 00:11:32.278 "assigned_rate_limits": { 00:11:32.278 "rw_ios_per_sec": 0, 00:11:32.278 "rw_mbytes_per_sec": 0, 00:11:32.278 "r_mbytes_per_sec": 0, 00:11:32.278 "w_mbytes_per_sec": 0 00:11:32.278 }, 00:11:32.278 "claimed": true, 00:11:32.278 "claim_type": "exclusive_write", 00:11:32.278 "zoned": false, 00:11:32.278 "supported_io_types": { 00:11:32.278 "read": true, 00:11:32.278 "write": true, 00:11:32.278 "unmap": true, 00:11:32.278 "flush": true, 00:11:32.278 "reset": true, 00:11:32.278 "nvme_admin": false, 00:11:32.278 "nvme_io": false, 00:11:32.278 "nvme_io_md": false, 00:11:32.278 "write_zeroes": true, 00:11:32.278 "zcopy": true, 00:11:32.278 "get_zone_info": false, 00:11:32.278 "zone_management": false, 00:11:32.278 "zone_append": false, 00:11:32.278 "compare": false, 00:11:32.278 "compare_and_write": false, 00:11:32.278 "abort": true, 00:11:32.278 "seek_hole": false, 00:11:32.278 "seek_data": false, 00:11:32.278 "copy": true, 00:11:32.278 "nvme_iov_md": false 00:11:32.278 }, 00:11:32.278 "memory_domains": [ 00:11:32.278 { 00:11:32.278 "dma_device_id": "system", 00:11:32.278 "dma_device_type": 1 00:11:32.278 }, 00:11:32.278 { 00:11:32.278 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:32.278 "dma_device_type": 2 00:11:32.278 } 00:11:32.278 ], 00:11:32.278 "driver_specific": {} 00:11:32.278 } 00:11:32.278 ] 00:11:32.278 15:17:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:32.278 15:17:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:11:32.278 15:17:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:32.278 15:17:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:32.278 15:17:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:32.278 15:17:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:32.278 15:17:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:32.278 15:17:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:32.278 15:17:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:32.278 15:17:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:32.278 15:17:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:32.278 15:17:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:32.278 15:17:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:32.278 15:17:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:32.278 15:17:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:32.278 15:17:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:32.278 15:17:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:32.278 15:17:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:32.278 "name": "Existed_Raid", 00:11:32.278 "uuid": "90132b83-6ddd-45e9-b64e-2af0cdb7d01b", 00:11:32.278 "strip_size_kb": 64, 00:11:32.278 "state": "configuring", 00:11:32.278 "raid_level": "concat", 00:11:32.278 "superblock": true, 00:11:32.278 "num_base_bdevs": 4, 00:11:32.279 "num_base_bdevs_discovered": 3, 00:11:32.279 "num_base_bdevs_operational": 4, 00:11:32.279 "base_bdevs_list": [ 00:11:32.279 { 00:11:32.279 "name": "BaseBdev1", 00:11:32.279 "uuid": "5b3dd593-6129-4ec0-8d56-4355e82b10d9", 00:11:32.279 "is_configured": true, 00:11:32.279 "data_offset": 2048, 00:11:32.279 "data_size": 63488 00:11:32.279 }, 00:11:32.279 { 00:11:32.279 "name": null, 00:11:32.279 "uuid": "cdf7dc43-301f-450a-87ed-2e0c2bf0b35d", 00:11:32.279 "is_configured": false, 00:11:32.279 "data_offset": 0, 00:11:32.279 "data_size": 63488 00:11:32.279 }, 00:11:32.279 { 00:11:32.279 "name": "BaseBdev3", 00:11:32.279 "uuid": "0e75a818-d41e-45fd-9944-f503285583aa", 00:11:32.279 "is_configured": true, 00:11:32.279 "data_offset": 2048, 00:11:32.279 "data_size": 63488 00:11:32.279 }, 00:11:32.279 { 00:11:32.279 "name": "BaseBdev4", 00:11:32.279 "uuid": "8510c761-719a-4ea5-a504-27a8025e09ca", 00:11:32.279 "is_configured": true, 00:11:32.279 "data_offset": 2048, 00:11:32.279 "data_size": 63488 00:11:32.279 } 00:11:32.279 ] 00:11:32.279 }' 00:11:32.279 15:17:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:32.279 15:17:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:32.848 15:17:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:11:32.848 15:17:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:32.848 15:17:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:32.848 15:17:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:32.848 15:17:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:32.848 15:17:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:11:32.848 15:17:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:11:32.848 15:17:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:32.848 15:17:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:32.848 [2024-11-19 15:17:22.926214] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:11:32.848 15:17:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:32.848 15:17:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:32.848 15:17:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:32.848 15:17:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:32.848 15:17:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:32.848 15:17:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:32.848 15:17:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:32.848 15:17:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:32.848 15:17:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:32.848 15:17:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:32.848 15:17:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:32.848 15:17:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:32.848 15:17:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:32.848 15:17:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:32.848 15:17:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:32.848 15:17:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:32.848 15:17:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:32.848 "name": "Existed_Raid", 00:11:32.848 "uuid": "90132b83-6ddd-45e9-b64e-2af0cdb7d01b", 00:11:32.848 "strip_size_kb": 64, 00:11:32.848 "state": "configuring", 00:11:32.848 "raid_level": "concat", 00:11:32.848 "superblock": true, 00:11:32.848 "num_base_bdevs": 4, 00:11:32.848 "num_base_bdevs_discovered": 2, 00:11:32.848 "num_base_bdevs_operational": 4, 00:11:32.848 "base_bdevs_list": [ 00:11:32.848 { 00:11:32.848 "name": "BaseBdev1", 00:11:32.848 "uuid": "5b3dd593-6129-4ec0-8d56-4355e82b10d9", 00:11:32.848 "is_configured": true, 00:11:32.848 "data_offset": 2048, 00:11:32.848 "data_size": 63488 00:11:32.848 }, 00:11:32.848 { 00:11:32.848 "name": null, 00:11:32.848 "uuid": "cdf7dc43-301f-450a-87ed-2e0c2bf0b35d", 00:11:32.848 "is_configured": false, 00:11:32.848 "data_offset": 0, 00:11:32.849 "data_size": 63488 00:11:32.849 }, 00:11:32.849 { 00:11:32.849 "name": null, 00:11:32.849 "uuid": "0e75a818-d41e-45fd-9944-f503285583aa", 00:11:32.849 "is_configured": false, 00:11:32.849 "data_offset": 0, 00:11:32.849 "data_size": 63488 00:11:32.849 }, 00:11:32.849 { 00:11:32.849 "name": "BaseBdev4", 00:11:32.849 "uuid": "8510c761-719a-4ea5-a504-27a8025e09ca", 00:11:32.849 "is_configured": true, 00:11:32.849 "data_offset": 2048, 00:11:32.849 "data_size": 63488 00:11:32.849 } 00:11:32.849 ] 00:11:32.849 }' 00:11:32.849 15:17:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:32.849 15:17:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:33.108 15:17:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:33.108 15:17:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:33.108 15:17:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:33.108 15:17:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:11:33.108 15:17:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:33.366 15:17:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:11:33.366 15:17:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:11:33.366 15:17:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:33.366 15:17:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:33.366 [2024-11-19 15:17:23.453313] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:33.366 15:17:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:33.366 15:17:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:33.366 15:17:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:33.366 15:17:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:33.366 15:17:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:33.366 15:17:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:33.366 15:17:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:33.366 15:17:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:33.366 15:17:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:33.366 15:17:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:33.366 15:17:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:33.366 15:17:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:33.366 15:17:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:33.366 15:17:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:33.366 15:17:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:33.366 15:17:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:33.366 15:17:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:33.366 "name": "Existed_Raid", 00:11:33.366 "uuid": "90132b83-6ddd-45e9-b64e-2af0cdb7d01b", 00:11:33.366 "strip_size_kb": 64, 00:11:33.366 "state": "configuring", 00:11:33.367 "raid_level": "concat", 00:11:33.367 "superblock": true, 00:11:33.367 "num_base_bdevs": 4, 00:11:33.367 "num_base_bdevs_discovered": 3, 00:11:33.367 "num_base_bdevs_operational": 4, 00:11:33.367 "base_bdevs_list": [ 00:11:33.367 { 00:11:33.367 "name": "BaseBdev1", 00:11:33.367 "uuid": "5b3dd593-6129-4ec0-8d56-4355e82b10d9", 00:11:33.367 "is_configured": true, 00:11:33.367 "data_offset": 2048, 00:11:33.367 "data_size": 63488 00:11:33.367 }, 00:11:33.367 { 00:11:33.367 "name": null, 00:11:33.367 "uuid": "cdf7dc43-301f-450a-87ed-2e0c2bf0b35d", 00:11:33.367 "is_configured": false, 00:11:33.367 "data_offset": 0, 00:11:33.367 "data_size": 63488 00:11:33.367 }, 00:11:33.367 { 00:11:33.367 "name": "BaseBdev3", 00:11:33.367 "uuid": "0e75a818-d41e-45fd-9944-f503285583aa", 00:11:33.367 "is_configured": true, 00:11:33.367 "data_offset": 2048, 00:11:33.367 "data_size": 63488 00:11:33.367 }, 00:11:33.367 { 00:11:33.367 "name": "BaseBdev4", 00:11:33.367 "uuid": "8510c761-719a-4ea5-a504-27a8025e09ca", 00:11:33.367 "is_configured": true, 00:11:33.367 "data_offset": 2048, 00:11:33.367 "data_size": 63488 00:11:33.367 } 00:11:33.367 ] 00:11:33.367 }' 00:11:33.367 15:17:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:33.367 15:17:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:33.626 15:17:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:33.626 15:17:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:33.626 15:17:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:33.626 15:17:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:11:33.626 15:17:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:33.626 15:17:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:11:33.626 15:17:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:11:33.626 15:17:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:33.626 15:17:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:33.626 [2024-11-19 15:17:23.948565] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:33.885 15:17:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:33.885 15:17:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:33.885 15:17:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:33.885 15:17:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:33.885 15:17:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:33.885 15:17:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:33.885 15:17:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:33.885 15:17:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:33.885 15:17:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:33.885 15:17:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:33.885 15:17:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:33.885 15:17:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:33.885 15:17:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:33.885 15:17:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:33.885 15:17:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:33.885 15:17:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:33.885 15:17:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:33.885 "name": "Existed_Raid", 00:11:33.885 "uuid": "90132b83-6ddd-45e9-b64e-2af0cdb7d01b", 00:11:33.885 "strip_size_kb": 64, 00:11:33.885 "state": "configuring", 00:11:33.885 "raid_level": "concat", 00:11:33.885 "superblock": true, 00:11:33.885 "num_base_bdevs": 4, 00:11:33.885 "num_base_bdevs_discovered": 2, 00:11:33.885 "num_base_bdevs_operational": 4, 00:11:33.885 "base_bdevs_list": [ 00:11:33.885 { 00:11:33.885 "name": null, 00:11:33.885 "uuid": "5b3dd593-6129-4ec0-8d56-4355e82b10d9", 00:11:33.885 "is_configured": false, 00:11:33.885 "data_offset": 0, 00:11:33.885 "data_size": 63488 00:11:33.885 }, 00:11:33.885 { 00:11:33.885 "name": null, 00:11:33.885 "uuid": "cdf7dc43-301f-450a-87ed-2e0c2bf0b35d", 00:11:33.885 "is_configured": false, 00:11:33.885 "data_offset": 0, 00:11:33.885 "data_size": 63488 00:11:33.885 }, 00:11:33.885 { 00:11:33.885 "name": "BaseBdev3", 00:11:33.885 "uuid": "0e75a818-d41e-45fd-9944-f503285583aa", 00:11:33.885 "is_configured": true, 00:11:33.885 "data_offset": 2048, 00:11:33.885 "data_size": 63488 00:11:33.885 }, 00:11:33.885 { 00:11:33.885 "name": "BaseBdev4", 00:11:33.885 "uuid": "8510c761-719a-4ea5-a504-27a8025e09ca", 00:11:33.885 "is_configured": true, 00:11:33.885 "data_offset": 2048, 00:11:33.885 "data_size": 63488 00:11:33.885 } 00:11:33.885 ] 00:11:33.885 }' 00:11:33.885 15:17:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:33.885 15:17:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:34.144 15:17:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:34.144 15:17:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:11:34.144 15:17:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:34.144 15:17:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:34.144 15:17:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:34.144 15:17:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:11:34.144 15:17:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:11:34.144 15:17:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:34.144 15:17:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:34.144 [2024-11-19 15:17:24.455884] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:34.144 15:17:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:34.144 15:17:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:11:34.144 15:17:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:34.144 15:17:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:34.144 15:17:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:34.144 15:17:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:34.144 15:17:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:34.144 15:17:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:34.144 15:17:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:34.144 15:17:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:34.144 15:17:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:34.144 15:17:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:34.144 15:17:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:34.144 15:17:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:34.144 15:17:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:34.404 15:17:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:34.404 15:17:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:34.404 "name": "Existed_Raid", 00:11:34.404 "uuid": "90132b83-6ddd-45e9-b64e-2af0cdb7d01b", 00:11:34.404 "strip_size_kb": 64, 00:11:34.404 "state": "configuring", 00:11:34.404 "raid_level": "concat", 00:11:34.404 "superblock": true, 00:11:34.404 "num_base_bdevs": 4, 00:11:34.404 "num_base_bdevs_discovered": 3, 00:11:34.404 "num_base_bdevs_operational": 4, 00:11:34.404 "base_bdevs_list": [ 00:11:34.404 { 00:11:34.404 "name": null, 00:11:34.404 "uuid": "5b3dd593-6129-4ec0-8d56-4355e82b10d9", 00:11:34.404 "is_configured": false, 00:11:34.404 "data_offset": 0, 00:11:34.404 "data_size": 63488 00:11:34.404 }, 00:11:34.404 { 00:11:34.404 "name": "BaseBdev2", 00:11:34.404 "uuid": "cdf7dc43-301f-450a-87ed-2e0c2bf0b35d", 00:11:34.404 "is_configured": true, 00:11:34.404 "data_offset": 2048, 00:11:34.404 "data_size": 63488 00:11:34.404 }, 00:11:34.404 { 00:11:34.404 "name": "BaseBdev3", 00:11:34.404 "uuid": "0e75a818-d41e-45fd-9944-f503285583aa", 00:11:34.404 "is_configured": true, 00:11:34.404 "data_offset": 2048, 00:11:34.404 "data_size": 63488 00:11:34.404 }, 00:11:34.404 { 00:11:34.404 "name": "BaseBdev4", 00:11:34.404 "uuid": "8510c761-719a-4ea5-a504-27a8025e09ca", 00:11:34.404 "is_configured": true, 00:11:34.404 "data_offset": 2048, 00:11:34.404 "data_size": 63488 00:11:34.404 } 00:11:34.404 ] 00:11:34.404 }' 00:11:34.404 15:17:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:34.404 15:17:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:34.663 15:17:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:34.663 15:17:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:34.663 15:17:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:34.663 15:17:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:11:34.663 15:17:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:34.663 15:17:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:11:34.663 15:17:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:34.663 15:17:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:34.663 15:17:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:34.663 15:17:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:11:34.663 15:17:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:34.663 15:17:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 5b3dd593-6129-4ec0-8d56-4355e82b10d9 00:11:34.663 15:17:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:34.663 15:17:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:34.922 [2024-11-19 15:17:25.016056] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:11:34.922 [2024-11-19 15:17:25.016277] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:11:34.922 [2024-11-19 15:17:25.016290] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:11:34.922 [2024-11-19 15:17:25.016586] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:11:34.922 NewBaseBdev 00:11:34.922 [2024-11-19 15:17:25.016713] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:11:34.922 [2024-11-19 15:17:25.016724] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:11:34.922 [2024-11-19 15:17:25.016830] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:34.922 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:34.922 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:11:34.922 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:11:34.922 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:34.922 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:11:34.922 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:34.922 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:34.922 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:34.922 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:34.922 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:34.922 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:34.922 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:11:34.922 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:34.922 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:34.922 [ 00:11:34.922 { 00:11:34.922 "name": "NewBaseBdev", 00:11:34.922 "aliases": [ 00:11:34.922 "5b3dd593-6129-4ec0-8d56-4355e82b10d9" 00:11:34.922 ], 00:11:34.922 "product_name": "Malloc disk", 00:11:34.922 "block_size": 512, 00:11:34.922 "num_blocks": 65536, 00:11:34.922 "uuid": "5b3dd593-6129-4ec0-8d56-4355e82b10d9", 00:11:34.922 "assigned_rate_limits": { 00:11:34.922 "rw_ios_per_sec": 0, 00:11:34.922 "rw_mbytes_per_sec": 0, 00:11:34.922 "r_mbytes_per_sec": 0, 00:11:34.922 "w_mbytes_per_sec": 0 00:11:34.922 }, 00:11:34.922 "claimed": true, 00:11:34.922 "claim_type": "exclusive_write", 00:11:34.922 "zoned": false, 00:11:34.922 "supported_io_types": { 00:11:34.922 "read": true, 00:11:34.922 "write": true, 00:11:34.922 "unmap": true, 00:11:34.922 "flush": true, 00:11:34.922 "reset": true, 00:11:34.922 "nvme_admin": false, 00:11:34.922 "nvme_io": false, 00:11:34.922 "nvme_io_md": false, 00:11:34.922 "write_zeroes": true, 00:11:34.922 "zcopy": true, 00:11:34.922 "get_zone_info": false, 00:11:34.922 "zone_management": false, 00:11:34.922 "zone_append": false, 00:11:34.922 "compare": false, 00:11:34.922 "compare_and_write": false, 00:11:34.922 "abort": true, 00:11:34.922 "seek_hole": false, 00:11:34.922 "seek_data": false, 00:11:34.922 "copy": true, 00:11:34.922 "nvme_iov_md": false 00:11:34.922 }, 00:11:34.922 "memory_domains": [ 00:11:34.922 { 00:11:34.922 "dma_device_id": "system", 00:11:34.922 "dma_device_type": 1 00:11:34.922 }, 00:11:34.922 { 00:11:34.922 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:34.922 "dma_device_type": 2 00:11:34.922 } 00:11:34.922 ], 00:11:34.922 "driver_specific": {} 00:11:34.922 } 00:11:34.922 ] 00:11:34.922 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:34.922 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:11:34.922 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:11:34.922 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:34.922 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:34.922 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:34.922 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:34.922 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:34.922 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:34.922 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:34.922 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:34.922 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:34.922 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:34.922 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:34.922 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:34.922 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:34.922 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:34.923 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:34.923 "name": "Existed_Raid", 00:11:34.923 "uuid": "90132b83-6ddd-45e9-b64e-2af0cdb7d01b", 00:11:34.923 "strip_size_kb": 64, 00:11:34.923 "state": "online", 00:11:34.923 "raid_level": "concat", 00:11:34.923 "superblock": true, 00:11:34.923 "num_base_bdevs": 4, 00:11:34.923 "num_base_bdevs_discovered": 4, 00:11:34.923 "num_base_bdevs_operational": 4, 00:11:34.923 "base_bdevs_list": [ 00:11:34.923 { 00:11:34.923 "name": "NewBaseBdev", 00:11:34.923 "uuid": "5b3dd593-6129-4ec0-8d56-4355e82b10d9", 00:11:34.923 "is_configured": true, 00:11:34.923 "data_offset": 2048, 00:11:34.923 "data_size": 63488 00:11:34.923 }, 00:11:34.923 { 00:11:34.923 "name": "BaseBdev2", 00:11:34.923 "uuid": "cdf7dc43-301f-450a-87ed-2e0c2bf0b35d", 00:11:34.923 "is_configured": true, 00:11:34.923 "data_offset": 2048, 00:11:34.923 "data_size": 63488 00:11:34.923 }, 00:11:34.923 { 00:11:34.923 "name": "BaseBdev3", 00:11:34.923 "uuid": "0e75a818-d41e-45fd-9944-f503285583aa", 00:11:34.923 "is_configured": true, 00:11:34.923 "data_offset": 2048, 00:11:34.923 "data_size": 63488 00:11:34.923 }, 00:11:34.923 { 00:11:34.923 "name": "BaseBdev4", 00:11:34.923 "uuid": "8510c761-719a-4ea5-a504-27a8025e09ca", 00:11:34.923 "is_configured": true, 00:11:34.923 "data_offset": 2048, 00:11:34.923 "data_size": 63488 00:11:34.923 } 00:11:34.923 ] 00:11:34.923 }' 00:11:34.923 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:34.923 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:35.182 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:11:35.182 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:11:35.182 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:11:35.182 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:11:35.182 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:11:35.182 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:11:35.182 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:11:35.182 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:11:35.182 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:35.182 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:35.441 [2024-11-19 15:17:25.523630] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:35.441 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:35.441 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:11:35.441 "name": "Existed_Raid", 00:11:35.441 "aliases": [ 00:11:35.441 "90132b83-6ddd-45e9-b64e-2af0cdb7d01b" 00:11:35.441 ], 00:11:35.442 "product_name": "Raid Volume", 00:11:35.442 "block_size": 512, 00:11:35.442 "num_blocks": 253952, 00:11:35.442 "uuid": "90132b83-6ddd-45e9-b64e-2af0cdb7d01b", 00:11:35.442 "assigned_rate_limits": { 00:11:35.442 "rw_ios_per_sec": 0, 00:11:35.442 "rw_mbytes_per_sec": 0, 00:11:35.442 "r_mbytes_per_sec": 0, 00:11:35.442 "w_mbytes_per_sec": 0 00:11:35.442 }, 00:11:35.442 "claimed": false, 00:11:35.442 "zoned": false, 00:11:35.442 "supported_io_types": { 00:11:35.442 "read": true, 00:11:35.442 "write": true, 00:11:35.442 "unmap": true, 00:11:35.442 "flush": true, 00:11:35.442 "reset": true, 00:11:35.442 "nvme_admin": false, 00:11:35.442 "nvme_io": false, 00:11:35.442 "nvme_io_md": false, 00:11:35.442 "write_zeroes": true, 00:11:35.442 "zcopy": false, 00:11:35.442 "get_zone_info": false, 00:11:35.442 "zone_management": false, 00:11:35.442 "zone_append": false, 00:11:35.442 "compare": false, 00:11:35.442 "compare_and_write": false, 00:11:35.442 "abort": false, 00:11:35.442 "seek_hole": false, 00:11:35.442 "seek_data": false, 00:11:35.442 "copy": false, 00:11:35.442 "nvme_iov_md": false 00:11:35.442 }, 00:11:35.442 "memory_domains": [ 00:11:35.442 { 00:11:35.442 "dma_device_id": "system", 00:11:35.442 "dma_device_type": 1 00:11:35.442 }, 00:11:35.442 { 00:11:35.442 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:35.442 "dma_device_type": 2 00:11:35.442 }, 00:11:35.442 { 00:11:35.442 "dma_device_id": "system", 00:11:35.442 "dma_device_type": 1 00:11:35.442 }, 00:11:35.442 { 00:11:35.442 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:35.442 "dma_device_type": 2 00:11:35.442 }, 00:11:35.442 { 00:11:35.442 "dma_device_id": "system", 00:11:35.442 "dma_device_type": 1 00:11:35.442 }, 00:11:35.442 { 00:11:35.442 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:35.442 "dma_device_type": 2 00:11:35.442 }, 00:11:35.442 { 00:11:35.442 "dma_device_id": "system", 00:11:35.442 "dma_device_type": 1 00:11:35.442 }, 00:11:35.442 { 00:11:35.442 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:35.442 "dma_device_type": 2 00:11:35.442 } 00:11:35.442 ], 00:11:35.442 "driver_specific": { 00:11:35.442 "raid": { 00:11:35.442 "uuid": "90132b83-6ddd-45e9-b64e-2af0cdb7d01b", 00:11:35.442 "strip_size_kb": 64, 00:11:35.442 "state": "online", 00:11:35.442 "raid_level": "concat", 00:11:35.442 "superblock": true, 00:11:35.442 "num_base_bdevs": 4, 00:11:35.442 "num_base_bdevs_discovered": 4, 00:11:35.442 "num_base_bdevs_operational": 4, 00:11:35.442 "base_bdevs_list": [ 00:11:35.442 { 00:11:35.442 "name": "NewBaseBdev", 00:11:35.442 "uuid": "5b3dd593-6129-4ec0-8d56-4355e82b10d9", 00:11:35.442 "is_configured": true, 00:11:35.442 "data_offset": 2048, 00:11:35.442 "data_size": 63488 00:11:35.442 }, 00:11:35.442 { 00:11:35.442 "name": "BaseBdev2", 00:11:35.442 "uuid": "cdf7dc43-301f-450a-87ed-2e0c2bf0b35d", 00:11:35.442 "is_configured": true, 00:11:35.442 "data_offset": 2048, 00:11:35.442 "data_size": 63488 00:11:35.442 }, 00:11:35.442 { 00:11:35.442 "name": "BaseBdev3", 00:11:35.442 "uuid": "0e75a818-d41e-45fd-9944-f503285583aa", 00:11:35.442 "is_configured": true, 00:11:35.442 "data_offset": 2048, 00:11:35.442 "data_size": 63488 00:11:35.442 }, 00:11:35.442 { 00:11:35.442 "name": "BaseBdev4", 00:11:35.442 "uuid": "8510c761-719a-4ea5-a504-27a8025e09ca", 00:11:35.442 "is_configured": true, 00:11:35.442 "data_offset": 2048, 00:11:35.442 "data_size": 63488 00:11:35.442 } 00:11:35.442 ] 00:11:35.442 } 00:11:35.442 } 00:11:35.442 }' 00:11:35.442 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:11:35.442 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:11:35.442 BaseBdev2 00:11:35.442 BaseBdev3 00:11:35.442 BaseBdev4' 00:11:35.442 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:35.442 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:11:35.442 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:35.442 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:11:35.442 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:35.442 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:35.442 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:35.442 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:35.442 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:35.442 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:35.442 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:35.442 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:11:35.442 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:35.442 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:35.442 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:35.442 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:35.442 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:35.442 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:35.442 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:35.442 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:11:35.442 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:35.442 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:35.442 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:35.703 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:35.703 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:35.703 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:35.703 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:35.703 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:11:35.703 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:35.703 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:35.703 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:35.703 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:35.703 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:35.703 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:35.703 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:11:35.703 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:35.703 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:35.703 [2024-11-19 15:17:25.874672] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:11:35.703 [2024-11-19 15:17:25.874791] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:35.703 [2024-11-19 15:17:25.874896] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:35.703 [2024-11-19 15:17:25.875005] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:35.703 [2024-11-19 15:17:25.875060] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:11:35.703 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:35.703 15:17:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 82834 00:11:35.703 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 82834 ']' 00:11:35.703 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 82834 00:11:35.703 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:11:35.703 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:11:35.703 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 82834 00:11:35.703 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:11:35.703 killing process with pid 82834 00:11:35.703 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:11:35.703 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 82834' 00:11:35.703 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 82834 00:11:35.703 [2024-11-19 15:17:25.927210] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:35.703 15:17:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 82834 00:11:35.703 [2024-11-19 15:17:26.002842] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:36.298 15:17:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:11:36.298 00:11:36.298 real 0m10.034s 00:11:36.298 user 0m16.828s 00:11:36.298 sys 0m2.199s 00:11:36.298 ************************************ 00:11:36.298 END TEST raid_state_function_test_sb 00:11:36.298 ************************************ 00:11:36.298 15:17:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:36.298 15:17:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:36.298 15:17:26 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test concat 4 00:11:36.298 15:17:26 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:11:36.298 15:17:26 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:36.298 15:17:26 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:36.298 ************************************ 00:11:36.298 START TEST raid_superblock_test 00:11:36.298 ************************************ 00:11:36.298 15:17:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test concat 4 00:11:36.298 15:17:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=concat 00:11:36.298 15:17:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:11:36.298 15:17:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:11:36.298 15:17:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:11:36.298 15:17:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:11:36.298 15:17:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:11:36.298 15:17:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:11:36.298 15:17:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:11:36.298 15:17:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:11:36.298 15:17:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:11:36.298 15:17:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:11:36.298 15:17:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:11:36.298 15:17:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:11:36.298 15:17:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' concat '!=' raid1 ']' 00:11:36.298 15:17:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:11:36.298 15:17:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:11:36.298 15:17:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=83489 00:11:36.298 15:17:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:11:36.298 15:17:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 83489 00:11:36.298 15:17:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 83489 ']' 00:11:36.298 15:17:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:36.298 15:17:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:36.298 15:17:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:36.298 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:36.298 15:17:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:36.298 15:17:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:36.298 [2024-11-19 15:17:26.487955] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:11:36.298 [2024-11-19 15:17:26.488234] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid83489 ] 00:11:36.579 [2024-11-19 15:17:26.647232] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:36.579 [2024-11-19 15:17:26.686833] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:36.579 [2024-11-19 15:17:26.763842] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:36.579 [2024-11-19 15:17:26.764000] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:37.149 malloc1 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:37.149 [2024-11-19 15:17:27.418134] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:11:37.149 [2024-11-19 15:17:27.418300] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:37.149 [2024-11-19 15:17:27.418346] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:11:37.149 [2024-11-19 15:17:27.418398] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:37.149 [2024-11-19 15:17:27.420837] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:37.149 [2024-11-19 15:17:27.420916] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:11:37.149 pt1 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:37.149 malloc2 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:37.149 [2024-11-19 15:17:27.452686] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:11:37.149 [2024-11-19 15:17:27.452799] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:37.149 [2024-11-19 15:17:27.452830] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:11:37.149 [2024-11-19 15:17:27.452864] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:37.149 [2024-11-19 15:17:27.455136] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:37.149 [2024-11-19 15:17:27.455207] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:11:37.149 pt2 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:37.149 malloc3 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:37.149 15:17:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:37.410 [2024-11-19 15:17:27.491227] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:11:37.410 [2024-11-19 15:17:27.491291] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:37.410 [2024-11-19 15:17:27.491313] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:11:37.410 [2024-11-19 15:17:27.491326] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:37.410 [2024-11-19 15:17:27.493684] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:37.410 [2024-11-19 15:17:27.493789] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:11:37.410 pt3 00:11:37.410 15:17:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:37.410 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:11:37.410 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:11:37.410 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:11:37.410 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:11:37.410 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:11:37.410 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:11:37.410 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:11:37.410 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:11:37.410 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:11:37.410 15:17:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:37.410 15:17:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:37.410 malloc4 00:11:37.410 15:17:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:37.410 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:11:37.410 15:17:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:37.410 15:17:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:37.410 [2024-11-19 15:17:27.536635] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:11:37.410 [2024-11-19 15:17:27.536758] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:37.410 [2024-11-19 15:17:27.536791] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:11:37.410 [2024-11-19 15:17:27.536827] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:37.410 [2024-11-19 15:17:27.539188] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:37.410 [2024-11-19 15:17:27.539271] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:11:37.410 pt4 00:11:37.410 15:17:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:37.410 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:11:37.410 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:11:37.410 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:11:37.410 15:17:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:37.410 15:17:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:37.410 [2024-11-19 15:17:27.548653] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:11:37.410 [2024-11-19 15:17:27.550795] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:11:37.410 [2024-11-19 15:17:27.550901] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:11:37.410 [2024-11-19 15:17:27.550977] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:11:37.410 [2024-11-19 15:17:27.551172] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:11:37.410 [2024-11-19 15:17:27.551226] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:11:37.410 [2024-11-19 15:17:27.551495] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:11:37.410 [2024-11-19 15:17:27.551683] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:11:37.410 [2024-11-19 15:17:27.551742] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:11:37.410 [2024-11-19 15:17:27.551917] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:37.410 15:17:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:37.410 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:11:37.410 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:37.410 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:37.410 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:37.410 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:37.410 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:37.410 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:37.410 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:37.410 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:37.410 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:37.410 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:37.410 15:17:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:37.411 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:37.411 15:17:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:37.411 15:17:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:37.411 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:37.411 "name": "raid_bdev1", 00:11:37.411 "uuid": "a66f7357-61af-4d69-b6c7-15727804e871", 00:11:37.411 "strip_size_kb": 64, 00:11:37.411 "state": "online", 00:11:37.411 "raid_level": "concat", 00:11:37.411 "superblock": true, 00:11:37.411 "num_base_bdevs": 4, 00:11:37.411 "num_base_bdevs_discovered": 4, 00:11:37.411 "num_base_bdevs_operational": 4, 00:11:37.411 "base_bdevs_list": [ 00:11:37.411 { 00:11:37.411 "name": "pt1", 00:11:37.411 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:37.411 "is_configured": true, 00:11:37.411 "data_offset": 2048, 00:11:37.411 "data_size": 63488 00:11:37.411 }, 00:11:37.411 { 00:11:37.411 "name": "pt2", 00:11:37.411 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:37.411 "is_configured": true, 00:11:37.411 "data_offset": 2048, 00:11:37.411 "data_size": 63488 00:11:37.411 }, 00:11:37.411 { 00:11:37.411 "name": "pt3", 00:11:37.411 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:37.411 "is_configured": true, 00:11:37.411 "data_offset": 2048, 00:11:37.411 "data_size": 63488 00:11:37.411 }, 00:11:37.411 { 00:11:37.411 "name": "pt4", 00:11:37.411 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:37.411 "is_configured": true, 00:11:37.411 "data_offset": 2048, 00:11:37.411 "data_size": 63488 00:11:37.411 } 00:11:37.411 ] 00:11:37.411 }' 00:11:37.411 15:17:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:37.411 15:17:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:37.980 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:11:37.980 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:11:37.980 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:11:37.980 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:11:37.980 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:11:37.980 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:11:37.980 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:11:37.980 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:37.980 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:37.980 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:37.980 [2024-11-19 15:17:28.028252] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:37.980 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:37.980 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:11:37.980 "name": "raid_bdev1", 00:11:37.980 "aliases": [ 00:11:37.980 "a66f7357-61af-4d69-b6c7-15727804e871" 00:11:37.980 ], 00:11:37.980 "product_name": "Raid Volume", 00:11:37.980 "block_size": 512, 00:11:37.980 "num_blocks": 253952, 00:11:37.980 "uuid": "a66f7357-61af-4d69-b6c7-15727804e871", 00:11:37.980 "assigned_rate_limits": { 00:11:37.980 "rw_ios_per_sec": 0, 00:11:37.980 "rw_mbytes_per_sec": 0, 00:11:37.980 "r_mbytes_per_sec": 0, 00:11:37.980 "w_mbytes_per_sec": 0 00:11:37.980 }, 00:11:37.980 "claimed": false, 00:11:37.980 "zoned": false, 00:11:37.980 "supported_io_types": { 00:11:37.980 "read": true, 00:11:37.980 "write": true, 00:11:37.980 "unmap": true, 00:11:37.980 "flush": true, 00:11:37.980 "reset": true, 00:11:37.980 "nvme_admin": false, 00:11:37.980 "nvme_io": false, 00:11:37.980 "nvme_io_md": false, 00:11:37.980 "write_zeroes": true, 00:11:37.980 "zcopy": false, 00:11:37.980 "get_zone_info": false, 00:11:37.980 "zone_management": false, 00:11:37.980 "zone_append": false, 00:11:37.980 "compare": false, 00:11:37.980 "compare_and_write": false, 00:11:37.980 "abort": false, 00:11:37.980 "seek_hole": false, 00:11:37.980 "seek_data": false, 00:11:37.980 "copy": false, 00:11:37.980 "nvme_iov_md": false 00:11:37.980 }, 00:11:37.980 "memory_domains": [ 00:11:37.980 { 00:11:37.980 "dma_device_id": "system", 00:11:37.980 "dma_device_type": 1 00:11:37.980 }, 00:11:37.980 { 00:11:37.980 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:37.980 "dma_device_type": 2 00:11:37.980 }, 00:11:37.980 { 00:11:37.980 "dma_device_id": "system", 00:11:37.980 "dma_device_type": 1 00:11:37.980 }, 00:11:37.980 { 00:11:37.980 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:37.980 "dma_device_type": 2 00:11:37.980 }, 00:11:37.980 { 00:11:37.980 "dma_device_id": "system", 00:11:37.980 "dma_device_type": 1 00:11:37.980 }, 00:11:37.980 { 00:11:37.980 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:37.980 "dma_device_type": 2 00:11:37.980 }, 00:11:37.980 { 00:11:37.980 "dma_device_id": "system", 00:11:37.980 "dma_device_type": 1 00:11:37.980 }, 00:11:37.980 { 00:11:37.980 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:37.980 "dma_device_type": 2 00:11:37.980 } 00:11:37.980 ], 00:11:37.980 "driver_specific": { 00:11:37.980 "raid": { 00:11:37.980 "uuid": "a66f7357-61af-4d69-b6c7-15727804e871", 00:11:37.980 "strip_size_kb": 64, 00:11:37.980 "state": "online", 00:11:37.980 "raid_level": "concat", 00:11:37.980 "superblock": true, 00:11:37.980 "num_base_bdevs": 4, 00:11:37.980 "num_base_bdevs_discovered": 4, 00:11:37.980 "num_base_bdevs_operational": 4, 00:11:37.980 "base_bdevs_list": [ 00:11:37.980 { 00:11:37.980 "name": "pt1", 00:11:37.980 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:37.980 "is_configured": true, 00:11:37.980 "data_offset": 2048, 00:11:37.980 "data_size": 63488 00:11:37.980 }, 00:11:37.980 { 00:11:37.980 "name": "pt2", 00:11:37.980 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:37.980 "is_configured": true, 00:11:37.980 "data_offset": 2048, 00:11:37.980 "data_size": 63488 00:11:37.980 }, 00:11:37.980 { 00:11:37.980 "name": "pt3", 00:11:37.980 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:37.980 "is_configured": true, 00:11:37.980 "data_offset": 2048, 00:11:37.980 "data_size": 63488 00:11:37.980 }, 00:11:37.980 { 00:11:37.980 "name": "pt4", 00:11:37.980 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:37.980 "is_configured": true, 00:11:37.980 "data_offset": 2048, 00:11:37.980 "data_size": 63488 00:11:37.980 } 00:11:37.980 ] 00:11:37.980 } 00:11:37.980 } 00:11:37.980 }' 00:11:37.980 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:11:37.980 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:11:37.980 pt2 00:11:37.980 pt3 00:11:37.980 pt4' 00:11:37.980 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:37.980 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:11:37.980 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:37.980 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:11:37.981 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:37.981 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:37.981 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:37.981 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:37.981 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:37.981 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:37.981 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:37.981 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:11:37.981 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:37.981 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:37.981 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:37.981 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:37.981 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:37.981 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:37.981 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:37.981 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:11:37.981 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:37.981 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:37.981 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:37.981 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:37.981 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:37.981 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:37.981 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:37.981 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:11:37.981 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:37.981 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:37.981 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:37.981 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:37.981 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:37.981 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:37.981 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:11:37.981 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:37.981 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:37.981 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:38.240 [2024-11-19 15:17:28.323560] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=a66f7357-61af-4d69-b6c7-15727804e871 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z a66f7357-61af-4d69-b6c7-15727804e871 ']' 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:38.240 [2024-11-19 15:17:28.367201] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:38.240 [2024-11-19 15:17:28.367242] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:38.240 [2024-11-19 15:17:28.367366] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:38.240 [2024-11-19 15:17:28.367457] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:38.240 [2024-11-19 15:17:28.367471] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:38.240 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:38.240 [2024-11-19 15:17:28.534983] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:11:38.240 [2024-11-19 15:17:28.537274] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:11:38.240 [2024-11-19 15:17:28.537370] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:11:38.240 [2024-11-19 15:17:28.537418] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:11:38.240 [2024-11-19 15:17:28.537507] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:11:38.240 [2024-11-19 15:17:28.537604] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:11:38.241 [2024-11-19 15:17:28.537652] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:11:38.241 [2024-11-19 15:17:28.537669] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:11:38.241 [2024-11-19 15:17:28.537685] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:38.241 [2024-11-19 15:17:28.537695] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:11:38.241 request: 00:11:38.241 { 00:11:38.241 "name": "raid_bdev1", 00:11:38.241 "raid_level": "concat", 00:11:38.241 "base_bdevs": [ 00:11:38.241 "malloc1", 00:11:38.241 "malloc2", 00:11:38.241 "malloc3", 00:11:38.241 "malloc4" 00:11:38.241 ], 00:11:38.241 "strip_size_kb": 64, 00:11:38.241 "superblock": false, 00:11:38.241 "method": "bdev_raid_create", 00:11:38.241 "req_id": 1 00:11:38.241 } 00:11:38.241 Got JSON-RPC error response 00:11:38.241 response: 00:11:38.241 { 00:11:38.241 "code": -17, 00:11:38.241 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:11:38.241 } 00:11:38.241 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:11:38.241 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:11:38.241 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:11:38.241 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:11:38.241 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:11:38.241 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:38.241 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:11:38.241 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:38.241 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:38.241 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:38.501 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:11:38.501 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:11:38.501 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:11:38.501 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:38.501 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:38.501 [2024-11-19 15:17:28.602822] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:11:38.501 [2024-11-19 15:17:28.602940] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:38.501 [2024-11-19 15:17:28.602992] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:11:38.501 [2024-11-19 15:17:28.603060] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:38.501 [2024-11-19 15:17:28.605624] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:38.501 [2024-11-19 15:17:28.605695] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:11:38.501 [2024-11-19 15:17:28.605806] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:11:38.501 [2024-11-19 15:17:28.605880] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:11:38.501 pt1 00:11:38.501 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:38.501 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 4 00:11:38.501 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:38.501 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:38.501 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:38.501 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:38.501 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:38.501 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:38.501 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:38.501 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:38.501 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:38.501 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:38.501 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:38.501 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:38.501 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:38.501 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:38.501 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:38.501 "name": "raid_bdev1", 00:11:38.501 "uuid": "a66f7357-61af-4d69-b6c7-15727804e871", 00:11:38.501 "strip_size_kb": 64, 00:11:38.501 "state": "configuring", 00:11:38.501 "raid_level": "concat", 00:11:38.501 "superblock": true, 00:11:38.501 "num_base_bdevs": 4, 00:11:38.501 "num_base_bdevs_discovered": 1, 00:11:38.501 "num_base_bdevs_operational": 4, 00:11:38.501 "base_bdevs_list": [ 00:11:38.501 { 00:11:38.501 "name": "pt1", 00:11:38.501 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:38.501 "is_configured": true, 00:11:38.501 "data_offset": 2048, 00:11:38.501 "data_size": 63488 00:11:38.501 }, 00:11:38.501 { 00:11:38.501 "name": null, 00:11:38.501 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:38.501 "is_configured": false, 00:11:38.501 "data_offset": 2048, 00:11:38.501 "data_size": 63488 00:11:38.501 }, 00:11:38.501 { 00:11:38.501 "name": null, 00:11:38.501 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:38.501 "is_configured": false, 00:11:38.501 "data_offset": 2048, 00:11:38.501 "data_size": 63488 00:11:38.501 }, 00:11:38.501 { 00:11:38.501 "name": null, 00:11:38.501 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:38.501 "is_configured": false, 00:11:38.501 "data_offset": 2048, 00:11:38.501 "data_size": 63488 00:11:38.501 } 00:11:38.501 ] 00:11:38.501 }' 00:11:38.501 15:17:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:38.501 15:17:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:38.761 15:17:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:11:38.761 15:17:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:11:38.761 15:17:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:38.761 15:17:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:38.761 [2024-11-19 15:17:29.078031] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:11:38.761 [2024-11-19 15:17:29.078136] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:38.761 [2024-11-19 15:17:29.078162] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:11:38.761 [2024-11-19 15:17:29.078172] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:38.761 [2024-11-19 15:17:29.078674] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:38.761 [2024-11-19 15:17:29.078691] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:11:38.761 [2024-11-19 15:17:29.078780] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:11:38.761 [2024-11-19 15:17:29.078803] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:11:38.761 pt2 00:11:38.761 15:17:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:38.761 15:17:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:11:38.761 15:17:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:38.761 15:17:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:38.761 [2024-11-19 15:17:29.089998] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:11:38.761 15:17:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:38.761 15:17:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 4 00:11:38.761 15:17:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:38.761 15:17:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:38.761 15:17:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:38.761 15:17:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:38.761 15:17:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:38.761 15:17:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:38.761 15:17:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:38.761 15:17:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:38.761 15:17:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:39.022 15:17:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:39.022 15:17:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:39.022 15:17:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:39.022 15:17:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:39.022 15:17:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:39.022 15:17:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:39.022 "name": "raid_bdev1", 00:11:39.022 "uuid": "a66f7357-61af-4d69-b6c7-15727804e871", 00:11:39.022 "strip_size_kb": 64, 00:11:39.022 "state": "configuring", 00:11:39.022 "raid_level": "concat", 00:11:39.022 "superblock": true, 00:11:39.022 "num_base_bdevs": 4, 00:11:39.022 "num_base_bdevs_discovered": 1, 00:11:39.022 "num_base_bdevs_operational": 4, 00:11:39.022 "base_bdevs_list": [ 00:11:39.022 { 00:11:39.022 "name": "pt1", 00:11:39.022 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:39.022 "is_configured": true, 00:11:39.022 "data_offset": 2048, 00:11:39.022 "data_size": 63488 00:11:39.022 }, 00:11:39.022 { 00:11:39.022 "name": null, 00:11:39.022 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:39.022 "is_configured": false, 00:11:39.022 "data_offset": 0, 00:11:39.022 "data_size": 63488 00:11:39.022 }, 00:11:39.022 { 00:11:39.022 "name": null, 00:11:39.022 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:39.022 "is_configured": false, 00:11:39.022 "data_offset": 2048, 00:11:39.022 "data_size": 63488 00:11:39.022 }, 00:11:39.022 { 00:11:39.022 "name": null, 00:11:39.022 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:39.022 "is_configured": false, 00:11:39.022 "data_offset": 2048, 00:11:39.022 "data_size": 63488 00:11:39.022 } 00:11:39.022 ] 00:11:39.022 }' 00:11:39.022 15:17:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:39.022 15:17:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:39.283 15:17:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:11:39.283 15:17:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:11:39.283 15:17:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:11:39.283 15:17:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:39.283 15:17:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:39.283 [2024-11-19 15:17:29.545234] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:11:39.283 [2024-11-19 15:17:29.545409] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:39.283 [2024-11-19 15:17:29.545457] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:11:39.283 [2024-11-19 15:17:29.545501] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:39.283 [2024-11-19 15:17:29.546017] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:39.283 [2024-11-19 15:17:29.546082] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:11:39.283 [2024-11-19 15:17:29.546192] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:11:39.283 [2024-11-19 15:17:29.546256] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:11:39.283 pt2 00:11:39.283 15:17:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:39.283 15:17:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:11:39.283 15:17:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:11:39.283 15:17:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:11:39.283 15:17:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:39.283 15:17:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:39.283 [2024-11-19 15:17:29.557131] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:11:39.283 [2024-11-19 15:17:29.557223] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:39.283 [2024-11-19 15:17:29.557256] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:11:39.283 [2024-11-19 15:17:29.557291] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:39.283 [2024-11-19 15:17:29.557724] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:39.283 [2024-11-19 15:17:29.557779] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:11:39.283 [2024-11-19 15:17:29.557865] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:11:39.283 [2024-11-19 15:17:29.557913] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:11:39.283 pt3 00:11:39.283 15:17:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:39.283 15:17:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:11:39.283 15:17:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:11:39.283 15:17:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:11:39.283 15:17:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:39.283 15:17:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:39.283 [2024-11-19 15:17:29.569101] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:11:39.283 [2024-11-19 15:17:29.569155] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:39.283 [2024-11-19 15:17:29.569170] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:11:39.283 [2024-11-19 15:17:29.569181] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:39.283 [2024-11-19 15:17:29.569494] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:39.283 [2024-11-19 15:17:29.569512] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:11:39.283 [2024-11-19 15:17:29.569563] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:11:39.283 [2024-11-19 15:17:29.569582] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:11:39.283 [2024-11-19 15:17:29.569683] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:11:39.283 [2024-11-19 15:17:29.569695] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:11:39.283 [2024-11-19 15:17:29.569948] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:11:39.283 [2024-11-19 15:17:29.570086] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:11:39.283 [2024-11-19 15:17:29.570095] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:11:39.283 [2024-11-19 15:17:29.570193] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:39.283 pt4 00:11:39.283 15:17:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:39.283 15:17:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:11:39.283 15:17:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:11:39.283 15:17:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:11:39.283 15:17:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:39.283 15:17:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:39.283 15:17:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:39.283 15:17:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:39.283 15:17:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:39.283 15:17:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:39.283 15:17:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:39.283 15:17:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:39.283 15:17:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:39.283 15:17:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:39.283 15:17:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:39.283 15:17:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:39.283 15:17:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:39.283 15:17:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:39.545 15:17:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:39.545 "name": "raid_bdev1", 00:11:39.545 "uuid": "a66f7357-61af-4d69-b6c7-15727804e871", 00:11:39.545 "strip_size_kb": 64, 00:11:39.545 "state": "online", 00:11:39.545 "raid_level": "concat", 00:11:39.545 "superblock": true, 00:11:39.545 "num_base_bdevs": 4, 00:11:39.545 "num_base_bdevs_discovered": 4, 00:11:39.545 "num_base_bdevs_operational": 4, 00:11:39.545 "base_bdevs_list": [ 00:11:39.545 { 00:11:39.545 "name": "pt1", 00:11:39.545 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:39.545 "is_configured": true, 00:11:39.545 "data_offset": 2048, 00:11:39.545 "data_size": 63488 00:11:39.545 }, 00:11:39.545 { 00:11:39.545 "name": "pt2", 00:11:39.545 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:39.545 "is_configured": true, 00:11:39.545 "data_offset": 2048, 00:11:39.545 "data_size": 63488 00:11:39.545 }, 00:11:39.545 { 00:11:39.545 "name": "pt3", 00:11:39.545 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:39.545 "is_configured": true, 00:11:39.545 "data_offset": 2048, 00:11:39.545 "data_size": 63488 00:11:39.545 }, 00:11:39.545 { 00:11:39.545 "name": "pt4", 00:11:39.545 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:39.545 "is_configured": true, 00:11:39.545 "data_offset": 2048, 00:11:39.545 "data_size": 63488 00:11:39.545 } 00:11:39.545 ] 00:11:39.545 }' 00:11:39.545 15:17:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:39.545 15:17:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:39.806 15:17:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:11:39.806 15:17:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:11:39.806 15:17:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:11:39.806 15:17:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:11:39.806 15:17:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:11:39.806 15:17:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:11:39.806 15:17:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:39.806 15:17:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:11:39.806 15:17:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:39.806 15:17:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:39.806 [2024-11-19 15:17:30.048783] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:39.806 15:17:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:39.806 15:17:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:11:39.806 "name": "raid_bdev1", 00:11:39.806 "aliases": [ 00:11:39.806 "a66f7357-61af-4d69-b6c7-15727804e871" 00:11:39.806 ], 00:11:39.806 "product_name": "Raid Volume", 00:11:39.806 "block_size": 512, 00:11:39.806 "num_blocks": 253952, 00:11:39.806 "uuid": "a66f7357-61af-4d69-b6c7-15727804e871", 00:11:39.806 "assigned_rate_limits": { 00:11:39.806 "rw_ios_per_sec": 0, 00:11:39.806 "rw_mbytes_per_sec": 0, 00:11:39.806 "r_mbytes_per_sec": 0, 00:11:39.806 "w_mbytes_per_sec": 0 00:11:39.806 }, 00:11:39.806 "claimed": false, 00:11:39.806 "zoned": false, 00:11:39.806 "supported_io_types": { 00:11:39.806 "read": true, 00:11:39.806 "write": true, 00:11:39.806 "unmap": true, 00:11:39.806 "flush": true, 00:11:39.806 "reset": true, 00:11:39.806 "nvme_admin": false, 00:11:39.806 "nvme_io": false, 00:11:39.806 "nvme_io_md": false, 00:11:39.806 "write_zeroes": true, 00:11:39.806 "zcopy": false, 00:11:39.806 "get_zone_info": false, 00:11:39.806 "zone_management": false, 00:11:39.806 "zone_append": false, 00:11:39.806 "compare": false, 00:11:39.806 "compare_and_write": false, 00:11:39.806 "abort": false, 00:11:39.806 "seek_hole": false, 00:11:39.806 "seek_data": false, 00:11:39.806 "copy": false, 00:11:39.806 "nvme_iov_md": false 00:11:39.806 }, 00:11:39.806 "memory_domains": [ 00:11:39.806 { 00:11:39.806 "dma_device_id": "system", 00:11:39.806 "dma_device_type": 1 00:11:39.806 }, 00:11:39.806 { 00:11:39.806 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:39.806 "dma_device_type": 2 00:11:39.806 }, 00:11:39.806 { 00:11:39.806 "dma_device_id": "system", 00:11:39.806 "dma_device_type": 1 00:11:39.806 }, 00:11:39.806 { 00:11:39.806 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:39.806 "dma_device_type": 2 00:11:39.806 }, 00:11:39.806 { 00:11:39.806 "dma_device_id": "system", 00:11:39.806 "dma_device_type": 1 00:11:39.806 }, 00:11:39.806 { 00:11:39.806 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:39.806 "dma_device_type": 2 00:11:39.806 }, 00:11:39.806 { 00:11:39.806 "dma_device_id": "system", 00:11:39.806 "dma_device_type": 1 00:11:39.806 }, 00:11:39.806 { 00:11:39.806 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:39.806 "dma_device_type": 2 00:11:39.806 } 00:11:39.806 ], 00:11:39.806 "driver_specific": { 00:11:39.806 "raid": { 00:11:39.806 "uuid": "a66f7357-61af-4d69-b6c7-15727804e871", 00:11:39.806 "strip_size_kb": 64, 00:11:39.806 "state": "online", 00:11:39.806 "raid_level": "concat", 00:11:39.806 "superblock": true, 00:11:39.806 "num_base_bdevs": 4, 00:11:39.806 "num_base_bdevs_discovered": 4, 00:11:39.806 "num_base_bdevs_operational": 4, 00:11:39.806 "base_bdevs_list": [ 00:11:39.806 { 00:11:39.806 "name": "pt1", 00:11:39.806 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:39.806 "is_configured": true, 00:11:39.806 "data_offset": 2048, 00:11:39.806 "data_size": 63488 00:11:39.806 }, 00:11:39.806 { 00:11:39.806 "name": "pt2", 00:11:39.806 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:39.806 "is_configured": true, 00:11:39.806 "data_offset": 2048, 00:11:39.806 "data_size": 63488 00:11:39.806 }, 00:11:39.806 { 00:11:39.806 "name": "pt3", 00:11:39.806 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:39.806 "is_configured": true, 00:11:39.806 "data_offset": 2048, 00:11:39.806 "data_size": 63488 00:11:39.806 }, 00:11:39.806 { 00:11:39.806 "name": "pt4", 00:11:39.806 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:39.806 "is_configured": true, 00:11:39.806 "data_offset": 2048, 00:11:39.806 "data_size": 63488 00:11:39.806 } 00:11:39.806 ] 00:11:39.806 } 00:11:39.806 } 00:11:39.806 }' 00:11:39.806 15:17:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:11:40.067 15:17:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:11:40.067 pt2 00:11:40.067 pt3 00:11:40.067 pt4' 00:11:40.067 15:17:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:40.067 15:17:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:11:40.067 15:17:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:40.067 15:17:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:11:40.067 15:17:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:40.067 15:17:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:40.067 15:17:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:40.067 15:17:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:40.067 15:17:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:40.067 15:17:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:40.067 15:17:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:40.067 15:17:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:11:40.067 15:17:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:40.067 15:17:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:40.067 15:17:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:40.067 15:17:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:40.067 15:17:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:40.067 15:17:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:40.067 15:17:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:40.067 15:17:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:11:40.067 15:17:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:40.067 15:17:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:40.067 15:17:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:40.067 15:17:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:40.067 15:17:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:40.067 15:17:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:40.067 15:17:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:40.067 15:17:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:11:40.067 15:17:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:40.067 15:17:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:40.067 15:17:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:40.067 15:17:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:40.067 15:17:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:40.067 15:17:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:40.068 15:17:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:40.068 15:17:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:11:40.068 15:17:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:40.068 15:17:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:40.068 [2024-11-19 15:17:30.404224] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:40.328 15:17:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:40.328 15:17:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' a66f7357-61af-4d69-b6c7-15727804e871 '!=' a66f7357-61af-4d69-b6c7-15727804e871 ']' 00:11:40.328 15:17:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy concat 00:11:40.328 15:17:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:11:40.328 15:17:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:11:40.328 15:17:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 83489 00:11:40.328 15:17:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 83489 ']' 00:11:40.328 15:17:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 83489 00:11:40.328 15:17:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:11:40.328 15:17:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:11:40.328 15:17:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 83489 00:11:40.328 15:17:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:11:40.328 15:17:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:11:40.328 15:17:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 83489' 00:11:40.328 killing process with pid 83489 00:11:40.328 15:17:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 83489 00:11:40.328 [2024-11-19 15:17:30.489547] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:40.328 [2024-11-19 15:17:30.489693] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:40.328 15:17:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 83489 00:11:40.328 [2024-11-19 15:17:30.489776] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:40.328 [2024-11-19 15:17:30.489792] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:11:40.328 [2024-11-19 15:17:30.570730] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:40.588 15:17:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:11:40.588 00:11:40.588 real 0m4.495s 00:11:40.588 user 0m6.916s 00:11:40.588 sys 0m1.076s 00:11:40.588 ************************************ 00:11:40.588 END TEST raid_superblock_test 00:11:40.588 ************************************ 00:11:40.588 15:17:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:40.588 15:17:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:40.848 15:17:30 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test concat 4 read 00:11:40.848 15:17:30 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:11:40.848 15:17:30 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:40.848 15:17:30 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:40.848 ************************************ 00:11:40.848 START TEST raid_read_error_test 00:11:40.848 ************************************ 00:11:40.848 15:17:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test concat 4 read 00:11:40.848 15:17:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:11:40.848 15:17:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:11:40.848 15:17:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:11:40.848 15:17:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:11:40.848 15:17:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:40.848 15:17:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:11:40.848 15:17:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:40.848 15:17:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:40.848 15:17:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:11:40.848 15:17:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:40.848 15:17:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:40.848 15:17:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:11:40.848 15:17:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:40.848 15:17:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:40.848 15:17:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:11:40.848 15:17:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:40.848 15:17:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:40.848 15:17:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:11:40.848 15:17:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:11:40.848 15:17:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:11:40.848 15:17:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:11:40.848 15:17:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:11:40.848 15:17:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:11:40.848 15:17:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:11:40.848 15:17:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:11:40.848 15:17:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:11:40.848 15:17:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:11:40.848 15:17:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:11:40.848 15:17:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.R1PnuSI0vG 00:11:40.848 15:17:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=83743 00:11:40.848 15:17:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:11:40.848 15:17:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 83743 00:11:40.848 15:17:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 83743 ']' 00:11:40.848 15:17:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:40.848 15:17:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:40.848 15:17:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:40.848 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:40.848 15:17:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:40.848 15:17:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:40.848 [2024-11-19 15:17:31.074230] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:11:40.848 [2024-11-19 15:17:31.074459] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid83743 ] 00:11:41.108 [2024-11-19 15:17:31.229266] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:41.108 [2024-11-19 15:17:31.267637] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:41.108 [2024-11-19 15:17:31.343733] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:41.108 [2024-11-19 15:17:31.343782] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:41.678 15:17:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:41.678 15:17:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:11:41.678 15:17:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:41.678 15:17:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:11:41.678 15:17:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:41.679 15:17:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:41.679 BaseBdev1_malloc 00:11:41.679 15:17:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:41.679 15:17:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:11:41.679 15:17:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:41.679 15:17:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:41.679 true 00:11:41.679 15:17:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:41.679 15:17:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:11:41.679 15:17:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:41.679 15:17:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:41.679 [2024-11-19 15:17:32.005332] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:11:41.679 [2024-11-19 15:17:32.005396] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:41.679 [2024-11-19 15:17:32.005416] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:11:41.679 [2024-11-19 15:17:32.005432] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:41.679 [2024-11-19 15:17:32.007794] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:41.679 [2024-11-19 15:17:32.007833] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:41.679 BaseBdev1 00:11:41.679 15:17:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:41.679 15:17:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:41.679 15:17:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:11:41.679 15:17:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:41.679 15:17:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:41.939 BaseBdev2_malloc 00:11:41.939 15:17:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:41.939 15:17:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:11:41.939 15:17:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:41.939 15:17:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:41.939 true 00:11:41.939 15:17:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:41.939 15:17:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:11:41.939 15:17:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:41.939 15:17:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:41.939 [2024-11-19 15:17:32.052023] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:11:41.939 [2024-11-19 15:17:32.052156] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:41.939 [2024-11-19 15:17:32.052181] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:11:41.939 [2024-11-19 15:17:32.052200] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:41.939 [2024-11-19 15:17:32.054550] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:41.939 [2024-11-19 15:17:32.054590] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:11:41.939 BaseBdev2 00:11:41.939 15:17:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:41.939 15:17:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:41.939 15:17:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:11:41.939 15:17:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:41.939 15:17:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:41.939 BaseBdev3_malloc 00:11:41.939 15:17:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:41.939 15:17:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:11:41.939 15:17:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:41.939 15:17:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:41.939 true 00:11:41.939 15:17:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:41.939 15:17:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:11:41.939 15:17:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:41.939 15:17:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:41.939 [2024-11-19 15:17:32.098648] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:11:41.939 [2024-11-19 15:17:32.098707] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:41.939 [2024-11-19 15:17:32.098726] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:11:41.939 [2024-11-19 15:17:32.098735] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:41.939 [2024-11-19 15:17:32.101153] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:41.939 [2024-11-19 15:17:32.101190] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:11:41.939 BaseBdev3 00:11:41.939 15:17:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:41.939 15:17:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:41.939 15:17:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:11:41.939 15:17:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:41.939 15:17:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:41.939 BaseBdev4_malloc 00:11:41.939 15:17:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:41.939 15:17:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:11:41.940 15:17:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:41.940 15:17:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:41.940 true 00:11:41.940 15:17:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:41.940 15:17:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:11:41.940 15:17:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:41.940 15:17:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:41.940 [2024-11-19 15:17:32.157247] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:11:41.940 [2024-11-19 15:17:32.157316] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:41.940 [2024-11-19 15:17:32.157341] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:11:41.940 [2024-11-19 15:17:32.157351] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:41.940 [2024-11-19 15:17:32.159822] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:41.940 [2024-11-19 15:17:32.159899] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:11:41.940 BaseBdev4 00:11:41.940 15:17:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:41.940 15:17:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:11:41.940 15:17:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:41.940 15:17:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:41.940 [2024-11-19 15:17:32.169293] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:41.940 [2024-11-19 15:17:32.171472] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:41.940 [2024-11-19 15:17:32.171595] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:41.940 [2024-11-19 15:17:32.171686] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:41.940 [2024-11-19 15:17:32.171927] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002000 00:11:41.940 [2024-11-19 15:17:32.171941] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:11:41.940 [2024-11-19 15:17:32.172237] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002ef0 00:11:41.940 [2024-11-19 15:17:32.172382] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002000 00:11:41.940 [2024-11-19 15:17:32.172394] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002000 00:11:41.940 [2024-11-19 15:17:32.172533] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:41.940 15:17:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:41.940 15:17:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:11:41.940 15:17:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:41.940 15:17:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:41.940 15:17:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:41.940 15:17:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:41.940 15:17:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:41.940 15:17:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:41.940 15:17:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:41.940 15:17:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:41.940 15:17:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:41.940 15:17:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:41.940 15:17:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:41.940 15:17:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:41.940 15:17:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:41.940 15:17:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:41.940 15:17:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:41.940 "name": "raid_bdev1", 00:11:41.940 "uuid": "ce335e91-d76a-4e3c-b7e4-573d68c2f962", 00:11:41.940 "strip_size_kb": 64, 00:11:41.940 "state": "online", 00:11:41.940 "raid_level": "concat", 00:11:41.940 "superblock": true, 00:11:41.940 "num_base_bdevs": 4, 00:11:41.940 "num_base_bdevs_discovered": 4, 00:11:41.940 "num_base_bdevs_operational": 4, 00:11:41.940 "base_bdevs_list": [ 00:11:41.940 { 00:11:41.940 "name": "BaseBdev1", 00:11:41.940 "uuid": "6c211282-7d03-53fc-a78b-5e5dc7d5426b", 00:11:41.940 "is_configured": true, 00:11:41.940 "data_offset": 2048, 00:11:41.940 "data_size": 63488 00:11:41.940 }, 00:11:41.940 { 00:11:41.940 "name": "BaseBdev2", 00:11:41.940 "uuid": "d1bd654c-ba5e-5655-9866-7e50fc903d17", 00:11:41.940 "is_configured": true, 00:11:41.940 "data_offset": 2048, 00:11:41.940 "data_size": 63488 00:11:41.940 }, 00:11:41.940 { 00:11:41.940 "name": "BaseBdev3", 00:11:41.940 "uuid": "0b4b15be-0f19-50ea-8590-0a4a3253aa22", 00:11:41.940 "is_configured": true, 00:11:41.940 "data_offset": 2048, 00:11:41.940 "data_size": 63488 00:11:41.940 }, 00:11:41.940 { 00:11:41.940 "name": "BaseBdev4", 00:11:41.940 "uuid": "5c65de9d-1645-50d4-992b-7971639f1289", 00:11:41.940 "is_configured": true, 00:11:41.940 "data_offset": 2048, 00:11:41.940 "data_size": 63488 00:11:41.940 } 00:11:41.940 ] 00:11:41.940 }' 00:11:41.940 15:17:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:41.940 15:17:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:42.510 15:17:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:11:42.510 15:17:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:11:42.510 [2024-11-19 15:17:32.752858] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000003090 00:11:43.447 15:17:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:11:43.447 15:17:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:43.447 15:17:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:43.447 15:17:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:43.447 15:17:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:11:43.447 15:17:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:11:43.447 15:17:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:11:43.447 15:17:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:11:43.447 15:17:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:43.447 15:17:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:43.447 15:17:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:43.447 15:17:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:43.447 15:17:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:43.447 15:17:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:43.447 15:17:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:43.447 15:17:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:43.447 15:17:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:43.448 15:17:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:43.448 15:17:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:43.448 15:17:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:43.448 15:17:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:43.448 15:17:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:43.448 15:17:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:43.448 "name": "raid_bdev1", 00:11:43.448 "uuid": "ce335e91-d76a-4e3c-b7e4-573d68c2f962", 00:11:43.448 "strip_size_kb": 64, 00:11:43.448 "state": "online", 00:11:43.448 "raid_level": "concat", 00:11:43.448 "superblock": true, 00:11:43.448 "num_base_bdevs": 4, 00:11:43.448 "num_base_bdevs_discovered": 4, 00:11:43.448 "num_base_bdevs_operational": 4, 00:11:43.448 "base_bdevs_list": [ 00:11:43.448 { 00:11:43.448 "name": "BaseBdev1", 00:11:43.448 "uuid": "6c211282-7d03-53fc-a78b-5e5dc7d5426b", 00:11:43.448 "is_configured": true, 00:11:43.448 "data_offset": 2048, 00:11:43.448 "data_size": 63488 00:11:43.448 }, 00:11:43.448 { 00:11:43.448 "name": "BaseBdev2", 00:11:43.448 "uuid": "d1bd654c-ba5e-5655-9866-7e50fc903d17", 00:11:43.448 "is_configured": true, 00:11:43.448 "data_offset": 2048, 00:11:43.448 "data_size": 63488 00:11:43.448 }, 00:11:43.448 { 00:11:43.448 "name": "BaseBdev3", 00:11:43.448 "uuid": "0b4b15be-0f19-50ea-8590-0a4a3253aa22", 00:11:43.448 "is_configured": true, 00:11:43.448 "data_offset": 2048, 00:11:43.448 "data_size": 63488 00:11:43.448 }, 00:11:43.448 { 00:11:43.448 "name": "BaseBdev4", 00:11:43.448 "uuid": "5c65de9d-1645-50d4-992b-7971639f1289", 00:11:43.448 "is_configured": true, 00:11:43.448 "data_offset": 2048, 00:11:43.448 "data_size": 63488 00:11:43.448 } 00:11:43.448 ] 00:11:43.448 }' 00:11:43.448 15:17:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:43.448 15:17:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:44.019 15:17:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:44.019 15:17:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:44.019 15:17:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:44.019 [2024-11-19 15:17:34.130522] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:44.019 [2024-11-19 15:17:34.130582] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:44.019 [2024-11-19 15:17:34.132910] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:44.019 [2024-11-19 15:17:34.133070] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:44.019 [2024-11-19 15:17:34.133134] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:44.019 [2024-11-19 15:17:34.133144] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state offline 00:11:44.019 { 00:11:44.019 "results": [ 00:11:44.019 { 00:11:44.019 "job": "raid_bdev1", 00:11:44.019 "core_mask": "0x1", 00:11:44.019 "workload": "randrw", 00:11:44.019 "percentage": 50, 00:11:44.019 "status": "finished", 00:11:44.019 "queue_depth": 1, 00:11:44.019 "io_size": 131072, 00:11:44.019 "runtime": 1.377921, 00:11:44.019 "iops": 13897.023123967194, 00:11:44.019 "mibps": 1737.1278904958992, 00:11:44.019 "io_failed": 1, 00:11:44.019 "io_timeout": 0, 00:11:44.019 "avg_latency_us": 101.19019152405167, 00:11:44.019 "min_latency_us": 25.823580786026202, 00:11:44.019 "max_latency_us": 1473.844541484716 00:11:44.019 } 00:11:44.019 ], 00:11:44.019 "core_count": 1 00:11:44.019 } 00:11:44.019 15:17:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:44.019 15:17:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 83743 00:11:44.019 15:17:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 83743 ']' 00:11:44.019 15:17:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 83743 00:11:44.019 15:17:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:11:44.019 15:17:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:11:44.019 15:17:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 83743 00:11:44.019 15:17:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:11:44.019 15:17:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:11:44.019 15:17:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 83743' 00:11:44.019 killing process with pid 83743 00:11:44.019 15:17:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 83743 00:11:44.019 [2024-11-19 15:17:34.177913] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:44.019 15:17:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 83743 00:11:44.019 [2024-11-19 15:17:34.244726] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:44.280 15:17:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.R1PnuSI0vG 00:11:44.280 15:17:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:11:44.280 15:17:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:11:44.280 15:17:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.73 00:11:44.280 15:17:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:11:44.280 15:17:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:11:44.280 15:17:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:11:44.280 15:17:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.73 != \0\.\0\0 ]] 00:11:44.280 00:11:44.280 real 0m3.609s 00:11:44.280 user 0m4.544s 00:11:44.280 sys 0m0.605s 00:11:44.280 ************************************ 00:11:44.280 END TEST raid_read_error_test 00:11:44.280 ************************************ 00:11:44.280 15:17:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:44.280 15:17:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:44.540 15:17:34 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test concat 4 write 00:11:44.540 15:17:34 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:11:44.540 15:17:34 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:44.540 15:17:34 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:44.540 ************************************ 00:11:44.540 START TEST raid_write_error_test 00:11:44.540 ************************************ 00:11:44.540 15:17:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test concat 4 write 00:11:44.540 15:17:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:11:44.540 15:17:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:11:44.540 15:17:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:11:44.540 15:17:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:11:44.540 15:17:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:44.540 15:17:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:11:44.540 15:17:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:44.540 15:17:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:44.540 15:17:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:11:44.540 15:17:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:44.540 15:17:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:44.540 15:17:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:11:44.540 15:17:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:44.540 15:17:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:44.540 15:17:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:11:44.540 15:17:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:44.540 15:17:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:44.540 15:17:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:11:44.540 15:17:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:11:44.540 15:17:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:11:44.540 15:17:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:11:44.540 15:17:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:11:44.540 15:17:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:11:44.540 15:17:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:11:44.540 15:17:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:11:44.540 15:17:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:11:44.540 15:17:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:11:44.540 15:17:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:11:44.540 15:17:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.qyBuYP1hrh 00:11:44.540 15:17:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=83873 00:11:44.540 15:17:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:11:44.540 15:17:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 83873 00:11:44.540 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:44.540 15:17:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 83873 ']' 00:11:44.540 15:17:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:44.540 15:17:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:44.540 15:17:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:44.540 15:17:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:44.540 15:17:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:44.540 [2024-11-19 15:17:34.748673] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:11:44.540 [2024-11-19 15:17:34.748803] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid83873 ] 00:11:44.800 [2024-11-19 15:17:34.905779] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:44.800 [2024-11-19 15:17:34.943336] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:44.800 [2024-11-19 15:17:35.019324] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:44.800 [2024-11-19 15:17:35.019372] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:45.371 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:45.371 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:11:45.371 15:17:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:45.371 15:17:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:11:45.371 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:45.371 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:45.371 BaseBdev1_malloc 00:11:45.371 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:45.371 15:17:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:11:45.371 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:45.371 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:45.371 true 00:11:45.371 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:45.371 15:17:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:11:45.371 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:45.371 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:45.371 [2024-11-19 15:17:35.629815] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:11:45.371 [2024-11-19 15:17:35.629881] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:45.371 [2024-11-19 15:17:35.629912] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:11:45.371 [2024-11-19 15:17:35.629921] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:45.371 [2024-11-19 15:17:35.632414] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:45.371 [2024-11-19 15:17:35.632453] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:45.371 BaseBdev1 00:11:45.371 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:45.371 15:17:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:45.371 15:17:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:11:45.371 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:45.371 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:45.371 BaseBdev2_malloc 00:11:45.371 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:45.371 15:17:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:11:45.371 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:45.371 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:45.371 true 00:11:45.371 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:45.371 15:17:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:11:45.371 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:45.371 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:45.371 [2024-11-19 15:17:35.676196] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:11:45.371 [2024-11-19 15:17:35.676242] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:45.371 [2024-11-19 15:17:35.676261] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:11:45.371 [2024-11-19 15:17:35.676278] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:45.371 [2024-11-19 15:17:35.678523] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:45.371 [2024-11-19 15:17:35.678560] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:11:45.371 BaseBdev2 00:11:45.371 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:45.371 15:17:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:45.371 15:17:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:11:45.371 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:45.371 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:45.371 BaseBdev3_malloc 00:11:45.371 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:45.371 15:17:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:11:45.371 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:45.371 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:45.631 true 00:11:45.631 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:45.631 15:17:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:11:45.631 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:45.631 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:45.631 [2024-11-19 15:17:35.722806] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:11:45.632 [2024-11-19 15:17:35.722873] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:45.632 [2024-11-19 15:17:35.722894] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:11:45.632 [2024-11-19 15:17:35.722903] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:45.632 [2024-11-19 15:17:35.725395] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:45.632 [2024-11-19 15:17:35.725434] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:11:45.632 BaseBdev3 00:11:45.632 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:45.632 15:17:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:45.632 15:17:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:11:45.632 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:45.632 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:45.632 BaseBdev4_malloc 00:11:45.632 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:45.632 15:17:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:11:45.632 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:45.632 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:45.632 true 00:11:45.632 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:45.632 15:17:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:11:45.632 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:45.632 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:45.632 [2024-11-19 15:17:35.780623] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:11:45.632 [2024-11-19 15:17:35.780706] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:45.632 [2024-11-19 15:17:35.780738] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:11:45.632 [2024-11-19 15:17:35.780748] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:45.632 [2024-11-19 15:17:35.783231] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:45.632 [2024-11-19 15:17:35.783368] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:11:45.632 BaseBdev4 00:11:45.632 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:45.632 15:17:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:11:45.632 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:45.632 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:45.632 [2024-11-19 15:17:35.792646] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:45.632 [2024-11-19 15:17:35.794766] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:45.632 [2024-11-19 15:17:35.794939] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:45.632 [2024-11-19 15:17:35.795019] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:45.632 [2024-11-19 15:17:35.795244] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002000 00:11:45.632 [2024-11-19 15:17:35.795262] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:11:45.632 [2024-11-19 15:17:35.795598] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002ef0 00:11:45.632 [2024-11-19 15:17:35.795810] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002000 00:11:45.632 [2024-11-19 15:17:35.795824] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002000 00:11:45.632 [2024-11-19 15:17:35.796033] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:45.632 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:45.632 15:17:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:11:45.632 15:17:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:45.632 15:17:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:45.632 15:17:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:45.632 15:17:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:45.632 15:17:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:45.632 15:17:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:45.632 15:17:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:45.632 15:17:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:45.632 15:17:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:45.632 15:17:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:45.632 15:17:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:45.632 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:45.632 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:45.632 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:45.632 15:17:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:45.632 "name": "raid_bdev1", 00:11:45.632 "uuid": "0cb34c40-7d8c-41d7-91f0-a8bdd2d0d9a7", 00:11:45.632 "strip_size_kb": 64, 00:11:45.632 "state": "online", 00:11:45.632 "raid_level": "concat", 00:11:45.632 "superblock": true, 00:11:45.632 "num_base_bdevs": 4, 00:11:45.632 "num_base_bdevs_discovered": 4, 00:11:45.632 "num_base_bdevs_operational": 4, 00:11:45.632 "base_bdevs_list": [ 00:11:45.632 { 00:11:45.632 "name": "BaseBdev1", 00:11:45.632 "uuid": "f81cc160-7c5d-55bb-a229-1fc08d59f804", 00:11:45.632 "is_configured": true, 00:11:45.632 "data_offset": 2048, 00:11:45.632 "data_size": 63488 00:11:45.632 }, 00:11:45.632 { 00:11:45.632 "name": "BaseBdev2", 00:11:45.632 "uuid": "8430f4e7-5ea1-5b80-beab-5e3d86a9dc70", 00:11:45.632 "is_configured": true, 00:11:45.632 "data_offset": 2048, 00:11:45.632 "data_size": 63488 00:11:45.632 }, 00:11:45.632 { 00:11:45.632 "name": "BaseBdev3", 00:11:45.632 "uuid": "628d668c-94b4-5194-a7ea-1650940f5269", 00:11:45.632 "is_configured": true, 00:11:45.632 "data_offset": 2048, 00:11:45.632 "data_size": 63488 00:11:45.632 }, 00:11:45.632 { 00:11:45.632 "name": "BaseBdev4", 00:11:45.632 "uuid": "73516d36-2e29-5d62-82a0-bb62f0ba3056", 00:11:45.632 "is_configured": true, 00:11:45.632 "data_offset": 2048, 00:11:45.632 "data_size": 63488 00:11:45.632 } 00:11:45.632 ] 00:11:45.632 }' 00:11:45.632 15:17:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:45.632 15:17:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:45.892 15:17:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:11:45.892 15:17:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:11:46.152 [2024-11-19 15:17:36.272390] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000003090 00:11:47.091 15:17:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:11:47.091 15:17:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:47.091 15:17:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:47.091 15:17:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:47.091 15:17:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:11:47.091 15:17:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:11:47.091 15:17:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:11:47.091 15:17:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:11:47.091 15:17:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:47.091 15:17:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:47.091 15:17:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:11:47.091 15:17:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:11:47.091 15:17:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:47.091 15:17:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:47.091 15:17:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:47.091 15:17:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:47.091 15:17:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:47.091 15:17:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:47.091 15:17:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:47.091 15:17:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:47.091 15:17:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:47.091 15:17:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:47.091 15:17:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:47.091 "name": "raid_bdev1", 00:11:47.091 "uuid": "0cb34c40-7d8c-41d7-91f0-a8bdd2d0d9a7", 00:11:47.091 "strip_size_kb": 64, 00:11:47.091 "state": "online", 00:11:47.091 "raid_level": "concat", 00:11:47.091 "superblock": true, 00:11:47.091 "num_base_bdevs": 4, 00:11:47.091 "num_base_bdevs_discovered": 4, 00:11:47.091 "num_base_bdevs_operational": 4, 00:11:47.091 "base_bdevs_list": [ 00:11:47.091 { 00:11:47.091 "name": "BaseBdev1", 00:11:47.091 "uuid": "f81cc160-7c5d-55bb-a229-1fc08d59f804", 00:11:47.092 "is_configured": true, 00:11:47.092 "data_offset": 2048, 00:11:47.092 "data_size": 63488 00:11:47.092 }, 00:11:47.092 { 00:11:47.092 "name": "BaseBdev2", 00:11:47.092 "uuid": "8430f4e7-5ea1-5b80-beab-5e3d86a9dc70", 00:11:47.092 "is_configured": true, 00:11:47.092 "data_offset": 2048, 00:11:47.092 "data_size": 63488 00:11:47.092 }, 00:11:47.092 { 00:11:47.092 "name": "BaseBdev3", 00:11:47.092 "uuid": "628d668c-94b4-5194-a7ea-1650940f5269", 00:11:47.092 "is_configured": true, 00:11:47.092 "data_offset": 2048, 00:11:47.092 "data_size": 63488 00:11:47.092 }, 00:11:47.092 { 00:11:47.092 "name": "BaseBdev4", 00:11:47.092 "uuid": "73516d36-2e29-5d62-82a0-bb62f0ba3056", 00:11:47.092 "is_configured": true, 00:11:47.092 "data_offset": 2048, 00:11:47.092 "data_size": 63488 00:11:47.092 } 00:11:47.092 ] 00:11:47.092 }' 00:11:47.092 15:17:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:47.092 15:17:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:47.352 15:17:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:47.352 15:17:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:47.352 15:17:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:47.352 [2024-11-19 15:17:37.661677] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:47.352 [2024-11-19 15:17:37.661815] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:47.352 [2024-11-19 15:17:37.664473] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:47.352 [2024-11-19 15:17:37.664545] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:47.352 [2024-11-19 15:17:37.664601] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:47.352 [2024-11-19 15:17:37.664612] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state offline 00:11:47.352 { 00:11:47.352 "results": [ 00:11:47.352 { 00:11:47.352 "job": "raid_bdev1", 00:11:47.352 "core_mask": "0x1", 00:11:47.352 "workload": "randrw", 00:11:47.352 "percentage": 50, 00:11:47.352 "status": "finished", 00:11:47.352 "queue_depth": 1, 00:11:47.352 "io_size": 131072, 00:11:47.352 "runtime": 1.389805, 00:11:47.352 "iops": 14087.587827069265, 00:11:47.352 "mibps": 1760.948478383658, 00:11:47.352 "io_failed": 1, 00:11:47.352 "io_timeout": 0, 00:11:47.352 "avg_latency_us": 99.95930398633308, 00:11:47.352 "min_latency_us": 24.929257641921396, 00:11:47.352 "max_latency_us": 1373.6803493449781 00:11:47.352 } 00:11:47.352 ], 00:11:47.352 "core_count": 1 00:11:47.352 } 00:11:47.352 15:17:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:47.352 15:17:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 83873 00:11:47.352 15:17:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 83873 ']' 00:11:47.352 15:17:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 83873 00:11:47.352 15:17:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:11:47.352 15:17:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:11:47.352 15:17:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 83873 00:11:47.612 15:17:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:11:47.612 15:17:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:11:47.612 15:17:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 83873' 00:11:47.612 killing process with pid 83873 00:11:47.612 15:17:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 83873 00:11:47.612 [2024-11-19 15:17:37.706038] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:47.612 15:17:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 83873 00:11:47.612 [2024-11-19 15:17:37.773330] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:47.872 15:17:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.qyBuYP1hrh 00:11:47.872 15:17:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:11:47.872 15:17:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:11:47.872 15:17:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.72 00:11:47.872 15:17:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:11:47.872 15:17:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:11:47.872 15:17:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:11:47.872 15:17:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.72 != \0\.\0\0 ]] 00:11:47.872 00:11:47.872 real 0m3.459s 00:11:47.872 user 0m4.197s 00:11:47.872 sys 0m0.634s 00:11:47.872 15:17:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:47.872 15:17:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:47.872 ************************************ 00:11:47.872 END TEST raid_write_error_test 00:11:47.872 ************************************ 00:11:47.872 15:17:38 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:11:47.872 15:17:38 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid1 4 false 00:11:47.872 15:17:38 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:11:47.872 15:17:38 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:47.872 15:17:38 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:47.872 ************************************ 00:11:47.872 START TEST raid_state_function_test 00:11:47.872 ************************************ 00:11:47.872 15:17:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 4 false 00:11:47.872 15:17:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:11:47.872 15:17:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:11:47.872 15:17:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:11:47.872 15:17:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:11:47.872 15:17:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:11:47.872 15:17:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:47.872 15:17:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:11:47.872 15:17:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:11:47.872 15:17:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:47.872 15:17:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:11:47.872 15:17:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:11:47.872 15:17:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:47.872 15:17:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:11:47.872 15:17:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:11:47.872 15:17:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:47.872 15:17:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:11:47.872 15:17:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:11:47.872 15:17:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:47.872 15:17:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:11:47.872 15:17:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:11:47.872 15:17:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:11:47.872 15:17:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:11:47.872 15:17:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:11:47.872 15:17:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:11:47.872 15:17:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:11:47.872 15:17:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:11:47.872 15:17:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:11:47.872 15:17:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:11:47.872 15:17:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=84011 00:11:47.872 15:17:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:11:47.872 15:17:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 84011' 00:11:47.872 Process raid pid: 84011 00:11:47.872 15:17:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 84011 00:11:47.872 15:17:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 84011 ']' 00:11:47.872 15:17:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:47.872 15:17:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:47.872 15:17:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:47.872 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:47.872 15:17:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:47.872 15:17:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:48.133 [2024-11-19 15:17:38.274905] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:11:48.133 [2024-11-19 15:17:38.275130] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:11:48.133 [2024-11-19 15:17:38.431365] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:48.393 [2024-11-19 15:17:38.472882] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:48.393 [2024-11-19 15:17:38.550339] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:48.393 [2024-11-19 15:17:38.550496] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:48.963 15:17:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:48.963 15:17:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:11:48.963 15:17:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:48.963 15:17:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:48.963 15:17:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:48.963 [2024-11-19 15:17:39.122311] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:11:48.964 [2024-11-19 15:17:39.122460] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:11:48.964 [2024-11-19 15:17:39.122474] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:48.964 [2024-11-19 15:17:39.122484] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:48.964 [2024-11-19 15:17:39.122490] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:11:48.964 [2024-11-19 15:17:39.122504] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:11:48.964 [2024-11-19 15:17:39.122509] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:11:48.964 [2024-11-19 15:17:39.122518] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:11:48.964 15:17:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:48.964 15:17:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:48.964 15:17:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:48.964 15:17:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:48.964 15:17:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:48.964 15:17:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:48.964 15:17:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:48.964 15:17:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:48.964 15:17:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:48.964 15:17:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:48.964 15:17:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:48.964 15:17:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:48.964 15:17:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:48.964 15:17:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:48.964 15:17:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:48.964 15:17:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:48.964 15:17:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:48.964 "name": "Existed_Raid", 00:11:48.964 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:48.964 "strip_size_kb": 0, 00:11:48.964 "state": "configuring", 00:11:48.964 "raid_level": "raid1", 00:11:48.964 "superblock": false, 00:11:48.964 "num_base_bdevs": 4, 00:11:48.964 "num_base_bdevs_discovered": 0, 00:11:48.964 "num_base_bdevs_operational": 4, 00:11:48.964 "base_bdevs_list": [ 00:11:48.964 { 00:11:48.964 "name": "BaseBdev1", 00:11:48.964 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:48.964 "is_configured": false, 00:11:48.964 "data_offset": 0, 00:11:48.964 "data_size": 0 00:11:48.964 }, 00:11:48.964 { 00:11:48.964 "name": "BaseBdev2", 00:11:48.964 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:48.964 "is_configured": false, 00:11:48.964 "data_offset": 0, 00:11:48.964 "data_size": 0 00:11:48.964 }, 00:11:48.964 { 00:11:48.964 "name": "BaseBdev3", 00:11:48.964 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:48.964 "is_configured": false, 00:11:48.964 "data_offset": 0, 00:11:48.964 "data_size": 0 00:11:48.964 }, 00:11:48.964 { 00:11:48.964 "name": "BaseBdev4", 00:11:48.964 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:48.964 "is_configured": false, 00:11:48.964 "data_offset": 0, 00:11:48.964 "data_size": 0 00:11:48.964 } 00:11:48.964 ] 00:11:48.964 }' 00:11:48.964 15:17:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:48.964 15:17:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:49.224 15:17:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:11:49.224 15:17:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:49.224 15:17:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:49.224 [2024-11-19 15:17:39.549488] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:11:49.224 [2024-11-19 15:17:39.549659] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:11:49.224 15:17:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:49.224 15:17:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:49.224 15:17:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:49.224 15:17:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:49.224 [2024-11-19 15:17:39.561422] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:11:49.224 [2024-11-19 15:17:39.561522] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:11:49.224 [2024-11-19 15:17:39.561566] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:49.224 [2024-11-19 15:17:39.561590] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:49.224 [2024-11-19 15:17:39.561608] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:11:49.224 [2024-11-19 15:17:39.561629] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:11:49.224 [2024-11-19 15:17:39.561646] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:11:49.224 [2024-11-19 15:17:39.561667] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:11:49.485 15:17:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:49.485 15:17:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:11:49.485 15:17:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:49.485 15:17:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:49.485 [2024-11-19 15:17:39.588681] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:49.485 BaseBdev1 00:11:49.485 15:17:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:49.485 15:17:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:11:49.485 15:17:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:11:49.485 15:17:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:49.485 15:17:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:11:49.485 15:17:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:49.485 15:17:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:49.485 15:17:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:49.485 15:17:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:49.485 15:17:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:49.485 15:17:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:49.485 15:17:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:11:49.485 15:17:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:49.485 15:17:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:49.485 [ 00:11:49.485 { 00:11:49.485 "name": "BaseBdev1", 00:11:49.485 "aliases": [ 00:11:49.485 "c10d7aa2-e187-42fc-995c-27e6ba81f4a8" 00:11:49.485 ], 00:11:49.485 "product_name": "Malloc disk", 00:11:49.485 "block_size": 512, 00:11:49.485 "num_blocks": 65536, 00:11:49.485 "uuid": "c10d7aa2-e187-42fc-995c-27e6ba81f4a8", 00:11:49.485 "assigned_rate_limits": { 00:11:49.485 "rw_ios_per_sec": 0, 00:11:49.485 "rw_mbytes_per_sec": 0, 00:11:49.485 "r_mbytes_per_sec": 0, 00:11:49.485 "w_mbytes_per_sec": 0 00:11:49.485 }, 00:11:49.485 "claimed": true, 00:11:49.485 "claim_type": "exclusive_write", 00:11:49.485 "zoned": false, 00:11:49.485 "supported_io_types": { 00:11:49.485 "read": true, 00:11:49.485 "write": true, 00:11:49.485 "unmap": true, 00:11:49.485 "flush": true, 00:11:49.485 "reset": true, 00:11:49.485 "nvme_admin": false, 00:11:49.485 "nvme_io": false, 00:11:49.485 "nvme_io_md": false, 00:11:49.485 "write_zeroes": true, 00:11:49.485 "zcopy": true, 00:11:49.485 "get_zone_info": false, 00:11:49.485 "zone_management": false, 00:11:49.485 "zone_append": false, 00:11:49.485 "compare": false, 00:11:49.485 "compare_and_write": false, 00:11:49.485 "abort": true, 00:11:49.485 "seek_hole": false, 00:11:49.485 "seek_data": false, 00:11:49.485 "copy": true, 00:11:49.485 "nvme_iov_md": false 00:11:49.485 }, 00:11:49.485 "memory_domains": [ 00:11:49.485 { 00:11:49.485 "dma_device_id": "system", 00:11:49.485 "dma_device_type": 1 00:11:49.485 }, 00:11:49.485 { 00:11:49.485 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:49.485 "dma_device_type": 2 00:11:49.485 } 00:11:49.485 ], 00:11:49.485 "driver_specific": {} 00:11:49.485 } 00:11:49.485 ] 00:11:49.485 15:17:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:49.485 15:17:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:11:49.485 15:17:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:49.485 15:17:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:49.485 15:17:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:49.485 15:17:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:49.485 15:17:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:49.485 15:17:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:49.485 15:17:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:49.485 15:17:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:49.485 15:17:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:49.485 15:17:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:49.485 15:17:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:49.485 15:17:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:49.485 15:17:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:49.485 15:17:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:49.485 15:17:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:49.485 15:17:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:49.485 "name": "Existed_Raid", 00:11:49.485 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:49.485 "strip_size_kb": 0, 00:11:49.485 "state": "configuring", 00:11:49.485 "raid_level": "raid1", 00:11:49.485 "superblock": false, 00:11:49.485 "num_base_bdevs": 4, 00:11:49.485 "num_base_bdevs_discovered": 1, 00:11:49.485 "num_base_bdevs_operational": 4, 00:11:49.485 "base_bdevs_list": [ 00:11:49.485 { 00:11:49.485 "name": "BaseBdev1", 00:11:49.485 "uuid": "c10d7aa2-e187-42fc-995c-27e6ba81f4a8", 00:11:49.485 "is_configured": true, 00:11:49.485 "data_offset": 0, 00:11:49.485 "data_size": 65536 00:11:49.485 }, 00:11:49.485 { 00:11:49.485 "name": "BaseBdev2", 00:11:49.485 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:49.485 "is_configured": false, 00:11:49.485 "data_offset": 0, 00:11:49.485 "data_size": 0 00:11:49.485 }, 00:11:49.485 { 00:11:49.485 "name": "BaseBdev3", 00:11:49.485 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:49.485 "is_configured": false, 00:11:49.485 "data_offset": 0, 00:11:49.485 "data_size": 0 00:11:49.485 }, 00:11:49.485 { 00:11:49.485 "name": "BaseBdev4", 00:11:49.485 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:49.485 "is_configured": false, 00:11:49.485 "data_offset": 0, 00:11:49.485 "data_size": 0 00:11:49.485 } 00:11:49.485 ] 00:11:49.485 }' 00:11:49.485 15:17:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:49.485 15:17:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:49.746 15:17:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:11:49.746 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:49.746 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:49.746 [2024-11-19 15:17:40.063984] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:11:49.746 [2024-11-19 15:17:40.064167] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:11:49.746 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:49.746 15:17:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:49.746 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:49.746 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:49.746 [2024-11-19 15:17:40.072033] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:49.746 [2024-11-19 15:17:40.074343] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:49.746 [2024-11-19 15:17:40.074428] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:49.746 [2024-11-19 15:17:40.074456] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:11:49.746 [2024-11-19 15:17:40.074478] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:11:49.746 [2024-11-19 15:17:40.074494] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:11:49.746 [2024-11-19 15:17:40.074513] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:11:49.746 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:49.746 15:17:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:11:49.746 15:17:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:49.746 15:17:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:49.746 15:17:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:49.746 15:17:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:49.746 15:17:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:49.746 15:17:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:49.746 15:17:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:49.746 15:17:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:49.746 15:17:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:49.746 15:17:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:49.746 15:17:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:49.746 15:17:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:49.746 15:17:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:49.746 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:49.746 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:50.073 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:50.074 15:17:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:50.074 "name": "Existed_Raid", 00:11:50.074 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:50.074 "strip_size_kb": 0, 00:11:50.074 "state": "configuring", 00:11:50.074 "raid_level": "raid1", 00:11:50.074 "superblock": false, 00:11:50.074 "num_base_bdevs": 4, 00:11:50.074 "num_base_bdevs_discovered": 1, 00:11:50.074 "num_base_bdevs_operational": 4, 00:11:50.074 "base_bdevs_list": [ 00:11:50.074 { 00:11:50.074 "name": "BaseBdev1", 00:11:50.074 "uuid": "c10d7aa2-e187-42fc-995c-27e6ba81f4a8", 00:11:50.074 "is_configured": true, 00:11:50.074 "data_offset": 0, 00:11:50.074 "data_size": 65536 00:11:50.074 }, 00:11:50.074 { 00:11:50.074 "name": "BaseBdev2", 00:11:50.074 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:50.074 "is_configured": false, 00:11:50.074 "data_offset": 0, 00:11:50.074 "data_size": 0 00:11:50.074 }, 00:11:50.074 { 00:11:50.074 "name": "BaseBdev3", 00:11:50.074 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:50.074 "is_configured": false, 00:11:50.074 "data_offset": 0, 00:11:50.074 "data_size": 0 00:11:50.074 }, 00:11:50.074 { 00:11:50.074 "name": "BaseBdev4", 00:11:50.074 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:50.074 "is_configured": false, 00:11:50.074 "data_offset": 0, 00:11:50.074 "data_size": 0 00:11:50.074 } 00:11:50.074 ] 00:11:50.074 }' 00:11:50.074 15:17:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:50.074 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:50.334 15:17:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:11:50.334 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:50.334 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:50.334 [2024-11-19 15:17:40.500812] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:50.334 BaseBdev2 00:11:50.334 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:50.334 15:17:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:11:50.334 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:11:50.334 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:50.334 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:11:50.334 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:50.334 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:50.334 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:50.334 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:50.334 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:50.334 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:50.334 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:11:50.334 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:50.334 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:50.334 [ 00:11:50.334 { 00:11:50.334 "name": "BaseBdev2", 00:11:50.334 "aliases": [ 00:11:50.334 "f784b173-69e5-4ed1-b4d1-cc69dcd00ce2" 00:11:50.334 ], 00:11:50.334 "product_name": "Malloc disk", 00:11:50.334 "block_size": 512, 00:11:50.334 "num_blocks": 65536, 00:11:50.334 "uuid": "f784b173-69e5-4ed1-b4d1-cc69dcd00ce2", 00:11:50.334 "assigned_rate_limits": { 00:11:50.334 "rw_ios_per_sec": 0, 00:11:50.334 "rw_mbytes_per_sec": 0, 00:11:50.334 "r_mbytes_per_sec": 0, 00:11:50.334 "w_mbytes_per_sec": 0 00:11:50.334 }, 00:11:50.334 "claimed": true, 00:11:50.334 "claim_type": "exclusive_write", 00:11:50.334 "zoned": false, 00:11:50.334 "supported_io_types": { 00:11:50.334 "read": true, 00:11:50.334 "write": true, 00:11:50.334 "unmap": true, 00:11:50.334 "flush": true, 00:11:50.334 "reset": true, 00:11:50.334 "nvme_admin": false, 00:11:50.334 "nvme_io": false, 00:11:50.334 "nvme_io_md": false, 00:11:50.334 "write_zeroes": true, 00:11:50.334 "zcopy": true, 00:11:50.334 "get_zone_info": false, 00:11:50.334 "zone_management": false, 00:11:50.334 "zone_append": false, 00:11:50.334 "compare": false, 00:11:50.334 "compare_and_write": false, 00:11:50.334 "abort": true, 00:11:50.334 "seek_hole": false, 00:11:50.334 "seek_data": false, 00:11:50.334 "copy": true, 00:11:50.334 "nvme_iov_md": false 00:11:50.334 }, 00:11:50.334 "memory_domains": [ 00:11:50.334 { 00:11:50.334 "dma_device_id": "system", 00:11:50.334 "dma_device_type": 1 00:11:50.334 }, 00:11:50.334 { 00:11:50.334 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:50.334 "dma_device_type": 2 00:11:50.334 } 00:11:50.334 ], 00:11:50.334 "driver_specific": {} 00:11:50.334 } 00:11:50.334 ] 00:11:50.334 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:50.334 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:11:50.334 15:17:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:11:50.334 15:17:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:50.334 15:17:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:50.334 15:17:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:50.334 15:17:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:50.334 15:17:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:50.334 15:17:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:50.334 15:17:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:50.334 15:17:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:50.334 15:17:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:50.335 15:17:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:50.335 15:17:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:50.335 15:17:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:50.335 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:50.335 15:17:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:50.335 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:50.335 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:50.335 15:17:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:50.335 "name": "Existed_Raid", 00:11:50.335 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:50.335 "strip_size_kb": 0, 00:11:50.335 "state": "configuring", 00:11:50.335 "raid_level": "raid1", 00:11:50.335 "superblock": false, 00:11:50.335 "num_base_bdevs": 4, 00:11:50.335 "num_base_bdevs_discovered": 2, 00:11:50.335 "num_base_bdevs_operational": 4, 00:11:50.335 "base_bdevs_list": [ 00:11:50.335 { 00:11:50.335 "name": "BaseBdev1", 00:11:50.335 "uuid": "c10d7aa2-e187-42fc-995c-27e6ba81f4a8", 00:11:50.335 "is_configured": true, 00:11:50.335 "data_offset": 0, 00:11:50.335 "data_size": 65536 00:11:50.335 }, 00:11:50.335 { 00:11:50.335 "name": "BaseBdev2", 00:11:50.335 "uuid": "f784b173-69e5-4ed1-b4d1-cc69dcd00ce2", 00:11:50.335 "is_configured": true, 00:11:50.335 "data_offset": 0, 00:11:50.335 "data_size": 65536 00:11:50.335 }, 00:11:50.335 { 00:11:50.335 "name": "BaseBdev3", 00:11:50.335 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:50.335 "is_configured": false, 00:11:50.335 "data_offset": 0, 00:11:50.335 "data_size": 0 00:11:50.335 }, 00:11:50.335 { 00:11:50.335 "name": "BaseBdev4", 00:11:50.335 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:50.335 "is_configured": false, 00:11:50.335 "data_offset": 0, 00:11:50.335 "data_size": 0 00:11:50.335 } 00:11:50.335 ] 00:11:50.335 }' 00:11:50.335 15:17:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:50.335 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:50.904 15:17:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:11:50.904 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:50.904 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:50.904 [2024-11-19 15:17:40.982255] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:50.904 BaseBdev3 00:11:50.904 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:50.904 15:17:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:11:50.904 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:11:50.904 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:50.904 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:11:50.904 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:50.904 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:50.904 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:50.904 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:50.904 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:50.904 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:50.904 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:11:50.904 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:50.904 15:17:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:50.904 [ 00:11:50.904 { 00:11:50.904 "name": "BaseBdev3", 00:11:50.904 "aliases": [ 00:11:50.904 "91c44a5f-7ce8-4207-bd40-3c2891366279" 00:11:50.904 ], 00:11:50.904 "product_name": "Malloc disk", 00:11:50.904 "block_size": 512, 00:11:50.904 "num_blocks": 65536, 00:11:50.904 "uuid": "91c44a5f-7ce8-4207-bd40-3c2891366279", 00:11:50.904 "assigned_rate_limits": { 00:11:50.905 "rw_ios_per_sec": 0, 00:11:50.905 "rw_mbytes_per_sec": 0, 00:11:50.905 "r_mbytes_per_sec": 0, 00:11:50.905 "w_mbytes_per_sec": 0 00:11:50.905 }, 00:11:50.905 "claimed": true, 00:11:50.905 "claim_type": "exclusive_write", 00:11:50.905 "zoned": false, 00:11:50.905 "supported_io_types": { 00:11:50.905 "read": true, 00:11:50.905 "write": true, 00:11:50.905 "unmap": true, 00:11:50.905 "flush": true, 00:11:50.905 "reset": true, 00:11:50.905 "nvme_admin": false, 00:11:50.905 "nvme_io": false, 00:11:50.905 "nvme_io_md": false, 00:11:50.905 "write_zeroes": true, 00:11:50.905 "zcopy": true, 00:11:50.905 "get_zone_info": false, 00:11:50.905 "zone_management": false, 00:11:50.905 "zone_append": false, 00:11:50.905 "compare": false, 00:11:50.905 "compare_and_write": false, 00:11:50.905 "abort": true, 00:11:50.905 "seek_hole": false, 00:11:50.905 "seek_data": false, 00:11:50.905 "copy": true, 00:11:50.905 "nvme_iov_md": false 00:11:50.905 }, 00:11:50.905 "memory_domains": [ 00:11:50.905 { 00:11:50.905 "dma_device_id": "system", 00:11:50.905 "dma_device_type": 1 00:11:50.905 }, 00:11:50.905 { 00:11:50.905 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:50.905 "dma_device_type": 2 00:11:50.905 } 00:11:50.905 ], 00:11:50.905 "driver_specific": {} 00:11:50.905 } 00:11:50.905 ] 00:11:50.905 15:17:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:50.905 15:17:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:11:50.905 15:17:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:11:50.905 15:17:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:50.905 15:17:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:50.905 15:17:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:50.905 15:17:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:50.905 15:17:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:50.905 15:17:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:50.905 15:17:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:50.905 15:17:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:50.905 15:17:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:50.905 15:17:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:50.905 15:17:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:50.905 15:17:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:50.905 15:17:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:50.905 15:17:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:50.905 15:17:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:50.905 15:17:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:50.905 15:17:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:50.905 "name": "Existed_Raid", 00:11:50.905 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:50.905 "strip_size_kb": 0, 00:11:50.905 "state": "configuring", 00:11:50.905 "raid_level": "raid1", 00:11:50.905 "superblock": false, 00:11:50.905 "num_base_bdevs": 4, 00:11:50.905 "num_base_bdevs_discovered": 3, 00:11:50.905 "num_base_bdevs_operational": 4, 00:11:50.905 "base_bdevs_list": [ 00:11:50.905 { 00:11:50.905 "name": "BaseBdev1", 00:11:50.905 "uuid": "c10d7aa2-e187-42fc-995c-27e6ba81f4a8", 00:11:50.905 "is_configured": true, 00:11:50.905 "data_offset": 0, 00:11:50.905 "data_size": 65536 00:11:50.905 }, 00:11:50.905 { 00:11:50.905 "name": "BaseBdev2", 00:11:50.905 "uuid": "f784b173-69e5-4ed1-b4d1-cc69dcd00ce2", 00:11:50.905 "is_configured": true, 00:11:50.905 "data_offset": 0, 00:11:50.905 "data_size": 65536 00:11:50.905 }, 00:11:50.905 { 00:11:50.905 "name": "BaseBdev3", 00:11:50.905 "uuid": "91c44a5f-7ce8-4207-bd40-3c2891366279", 00:11:50.905 "is_configured": true, 00:11:50.905 "data_offset": 0, 00:11:50.905 "data_size": 65536 00:11:50.905 }, 00:11:50.905 { 00:11:50.905 "name": "BaseBdev4", 00:11:50.905 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:50.905 "is_configured": false, 00:11:50.905 "data_offset": 0, 00:11:50.905 "data_size": 0 00:11:50.905 } 00:11:50.905 ] 00:11:50.905 }' 00:11:50.905 15:17:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:50.905 15:17:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:51.166 15:17:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:11:51.166 15:17:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:51.166 15:17:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:51.166 [2024-11-19 15:17:41.486317] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:51.166 [2024-11-19 15:17:41.486381] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:11:51.166 [2024-11-19 15:17:41.486390] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:11:51.166 [2024-11-19 15:17:41.486705] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:11:51.166 [2024-11-19 15:17:41.486868] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:11:51.166 [2024-11-19 15:17:41.486881] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:11:51.166 [2024-11-19 15:17:41.487125] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:51.166 BaseBdev4 00:11:51.166 15:17:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:51.166 15:17:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:11:51.166 15:17:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:11:51.166 15:17:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:51.166 15:17:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:11:51.166 15:17:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:51.166 15:17:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:51.166 15:17:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:51.166 15:17:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:51.166 15:17:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:51.166 15:17:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:51.166 15:17:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:11:51.166 15:17:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:51.166 15:17:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:51.426 [ 00:11:51.426 { 00:11:51.426 "name": "BaseBdev4", 00:11:51.426 "aliases": [ 00:11:51.426 "dfccc5b7-1a80-4e54-8c3b-c466d3013dfd" 00:11:51.426 ], 00:11:51.426 "product_name": "Malloc disk", 00:11:51.426 "block_size": 512, 00:11:51.426 "num_blocks": 65536, 00:11:51.426 "uuid": "dfccc5b7-1a80-4e54-8c3b-c466d3013dfd", 00:11:51.426 "assigned_rate_limits": { 00:11:51.426 "rw_ios_per_sec": 0, 00:11:51.426 "rw_mbytes_per_sec": 0, 00:11:51.426 "r_mbytes_per_sec": 0, 00:11:51.426 "w_mbytes_per_sec": 0 00:11:51.426 }, 00:11:51.426 "claimed": true, 00:11:51.426 "claim_type": "exclusive_write", 00:11:51.426 "zoned": false, 00:11:51.426 "supported_io_types": { 00:11:51.426 "read": true, 00:11:51.426 "write": true, 00:11:51.426 "unmap": true, 00:11:51.426 "flush": true, 00:11:51.426 "reset": true, 00:11:51.426 "nvme_admin": false, 00:11:51.426 "nvme_io": false, 00:11:51.426 "nvme_io_md": false, 00:11:51.426 "write_zeroes": true, 00:11:51.426 "zcopy": true, 00:11:51.426 "get_zone_info": false, 00:11:51.426 "zone_management": false, 00:11:51.426 "zone_append": false, 00:11:51.426 "compare": false, 00:11:51.426 "compare_and_write": false, 00:11:51.426 "abort": true, 00:11:51.426 "seek_hole": false, 00:11:51.426 "seek_data": false, 00:11:51.426 "copy": true, 00:11:51.426 "nvme_iov_md": false 00:11:51.426 }, 00:11:51.426 "memory_domains": [ 00:11:51.426 { 00:11:51.426 "dma_device_id": "system", 00:11:51.426 "dma_device_type": 1 00:11:51.426 }, 00:11:51.426 { 00:11:51.426 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:51.426 "dma_device_type": 2 00:11:51.426 } 00:11:51.426 ], 00:11:51.426 "driver_specific": {} 00:11:51.426 } 00:11:51.426 ] 00:11:51.426 15:17:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:51.426 15:17:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:11:51.427 15:17:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:11:51.427 15:17:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:51.427 15:17:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:11:51.427 15:17:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:51.427 15:17:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:51.427 15:17:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:51.427 15:17:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:51.427 15:17:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:51.427 15:17:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:51.427 15:17:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:51.427 15:17:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:51.427 15:17:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:51.427 15:17:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:51.427 15:17:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:51.427 15:17:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:51.427 15:17:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:51.427 15:17:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:51.427 15:17:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:51.427 "name": "Existed_Raid", 00:11:51.427 "uuid": "a1668b00-4e8c-4746-b186-30bdd09b7578", 00:11:51.427 "strip_size_kb": 0, 00:11:51.427 "state": "online", 00:11:51.427 "raid_level": "raid1", 00:11:51.427 "superblock": false, 00:11:51.427 "num_base_bdevs": 4, 00:11:51.427 "num_base_bdevs_discovered": 4, 00:11:51.427 "num_base_bdevs_operational": 4, 00:11:51.427 "base_bdevs_list": [ 00:11:51.427 { 00:11:51.427 "name": "BaseBdev1", 00:11:51.427 "uuid": "c10d7aa2-e187-42fc-995c-27e6ba81f4a8", 00:11:51.427 "is_configured": true, 00:11:51.427 "data_offset": 0, 00:11:51.427 "data_size": 65536 00:11:51.427 }, 00:11:51.427 { 00:11:51.427 "name": "BaseBdev2", 00:11:51.427 "uuid": "f784b173-69e5-4ed1-b4d1-cc69dcd00ce2", 00:11:51.427 "is_configured": true, 00:11:51.427 "data_offset": 0, 00:11:51.427 "data_size": 65536 00:11:51.427 }, 00:11:51.427 { 00:11:51.427 "name": "BaseBdev3", 00:11:51.427 "uuid": "91c44a5f-7ce8-4207-bd40-3c2891366279", 00:11:51.427 "is_configured": true, 00:11:51.427 "data_offset": 0, 00:11:51.427 "data_size": 65536 00:11:51.427 }, 00:11:51.427 { 00:11:51.427 "name": "BaseBdev4", 00:11:51.427 "uuid": "dfccc5b7-1a80-4e54-8c3b-c466d3013dfd", 00:11:51.427 "is_configured": true, 00:11:51.427 "data_offset": 0, 00:11:51.427 "data_size": 65536 00:11:51.427 } 00:11:51.427 ] 00:11:51.427 }' 00:11:51.427 15:17:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:51.427 15:17:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:51.687 15:17:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:11:51.687 15:17:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:11:51.687 15:17:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:11:51.687 15:17:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:11:51.687 15:17:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:11:51.687 15:17:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:11:51.687 15:17:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:11:51.687 15:17:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:51.687 15:17:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:11:51.687 15:17:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:51.687 [2024-11-19 15:17:41.954026] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:51.687 15:17:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:51.687 15:17:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:11:51.687 "name": "Existed_Raid", 00:11:51.687 "aliases": [ 00:11:51.687 "a1668b00-4e8c-4746-b186-30bdd09b7578" 00:11:51.687 ], 00:11:51.687 "product_name": "Raid Volume", 00:11:51.687 "block_size": 512, 00:11:51.687 "num_blocks": 65536, 00:11:51.687 "uuid": "a1668b00-4e8c-4746-b186-30bdd09b7578", 00:11:51.687 "assigned_rate_limits": { 00:11:51.687 "rw_ios_per_sec": 0, 00:11:51.687 "rw_mbytes_per_sec": 0, 00:11:51.687 "r_mbytes_per_sec": 0, 00:11:51.687 "w_mbytes_per_sec": 0 00:11:51.687 }, 00:11:51.687 "claimed": false, 00:11:51.687 "zoned": false, 00:11:51.687 "supported_io_types": { 00:11:51.687 "read": true, 00:11:51.687 "write": true, 00:11:51.687 "unmap": false, 00:11:51.687 "flush": false, 00:11:51.687 "reset": true, 00:11:51.687 "nvme_admin": false, 00:11:51.687 "nvme_io": false, 00:11:51.687 "nvme_io_md": false, 00:11:51.687 "write_zeroes": true, 00:11:51.687 "zcopy": false, 00:11:51.687 "get_zone_info": false, 00:11:51.687 "zone_management": false, 00:11:51.687 "zone_append": false, 00:11:51.687 "compare": false, 00:11:51.687 "compare_and_write": false, 00:11:51.687 "abort": false, 00:11:51.687 "seek_hole": false, 00:11:51.687 "seek_data": false, 00:11:51.687 "copy": false, 00:11:51.687 "nvme_iov_md": false 00:11:51.687 }, 00:11:51.687 "memory_domains": [ 00:11:51.687 { 00:11:51.687 "dma_device_id": "system", 00:11:51.687 "dma_device_type": 1 00:11:51.687 }, 00:11:51.687 { 00:11:51.687 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:51.687 "dma_device_type": 2 00:11:51.687 }, 00:11:51.687 { 00:11:51.687 "dma_device_id": "system", 00:11:51.687 "dma_device_type": 1 00:11:51.687 }, 00:11:51.687 { 00:11:51.687 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:51.687 "dma_device_type": 2 00:11:51.687 }, 00:11:51.687 { 00:11:51.687 "dma_device_id": "system", 00:11:51.687 "dma_device_type": 1 00:11:51.687 }, 00:11:51.687 { 00:11:51.687 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:51.687 "dma_device_type": 2 00:11:51.687 }, 00:11:51.687 { 00:11:51.687 "dma_device_id": "system", 00:11:51.687 "dma_device_type": 1 00:11:51.687 }, 00:11:51.687 { 00:11:51.687 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:51.687 "dma_device_type": 2 00:11:51.687 } 00:11:51.687 ], 00:11:51.687 "driver_specific": { 00:11:51.687 "raid": { 00:11:51.687 "uuid": "a1668b00-4e8c-4746-b186-30bdd09b7578", 00:11:51.687 "strip_size_kb": 0, 00:11:51.688 "state": "online", 00:11:51.688 "raid_level": "raid1", 00:11:51.688 "superblock": false, 00:11:51.688 "num_base_bdevs": 4, 00:11:51.688 "num_base_bdevs_discovered": 4, 00:11:51.688 "num_base_bdevs_operational": 4, 00:11:51.688 "base_bdevs_list": [ 00:11:51.688 { 00:11:51.688 "name": "BaseBdev1", 00:11:51.688 "uuid": "c10d7aa2-e187-42fc-995c-27e6ba81f4a8", 00:11:51.688 "is_configured": true, 00:11:51.688 "data_offset": 0, 00:11:51.688 "data_size": 65536 00:11:51.688 }, 00:11:51.688 { 00:11:51.688 "name": "BaseBdev2", 00:11:51.688 "uuid": "f784b173-69e5-4ed1-b4d1-cc69dcd00ce2", 00:11:51.688 "is_configured": true, 00:11:51.688 "data_offset": 0, 00:11:51.688 "data_size": 65536 00:11:51.688 }, 00:11:51.688 { 00:11:51.688 "name": "BaseBdev3", 00:11:51.688 "uuid": "91c44a5f-7ce8-4207-bd40-3c2891366279", 00:11:51.688 "is_configured": true, 00:11:51.688 "data_offset": 0, 00:11:51.688 "data_size": 65536 00:11:51.688 }, 00:11:51.688 { 00:11:51.688 "name": "BaseBdev4", 00:11:51.688 "uuid": "dfccc5b7-1a80-4e54-8c3b-c466d3013dfd", 00:11:51.688 "is_configured": true, 00:11:51.688 "data_offset": 0, 00:11:51.688 "data_size": 65536 00:11:51.688 } 00:11:51.688 ] 00:11:51.688 } 00:11:51.688 } 00:11:51.688 }' 00:11:51.688 15:17:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:11:51.688 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:11:51.688 BaseBdev2 00:11:51.688 BaseBdev3 00:11:51.688 BaseBdev4' 00:11:51.688 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:51.948 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:11:51.948 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:51.948 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:51.948 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:11:51.948 15:17:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:51.948 15:17:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:51.948 15:17:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:51.948 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:51.948 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:51.948 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:51.948 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:11:51.948 15:17:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:51.948 15:17:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:51.948 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:51.948 15:17:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:51.948 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:51.948 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:51.948 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:51.948 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:11:51.948 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:51.948 15:17:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:51.948 15:17:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:51.948 15:17:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:51.948 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:51.948 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:51.948 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:51.948 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:11:51.948 15:17:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:51.948 15:17:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:51.948 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:51.948 15:17:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:51.948 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:51.948 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:51.948 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:11:51.948 15:17:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:51.948 15:17:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:51.948 [2024-11-19 15:17:42.269150] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:52.209 15:17:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.209 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:11:52.209 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:11:52.209 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:11:52.209 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:11:52.209 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:11:52.210 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:11:52.210 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:52.210 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:52.210 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:52.210 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:52.210 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:11:52.210 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:52.210 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:52.210 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:52.210 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:52.210 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:52.210 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:52.210 15:17:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.210 15:17:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.210 15:17:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.210 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:52.210 "name": "Existed_Raid", 00:11:52.210 "uuid": "a1668b00-4e8c-4746-b186-30bdd09b7578", 00:11:52.210 "strip_size_kb": 0, 00:11:52.210 "state": "online", 00:11:52.210 "raid_level": "raid1", 00:11:52.210 "superblock": false, 00:11:52.210 "num_base_bdevs": 4, 00:11:52.210 "num_base_bdevs_discovered": 3, 00:11:52.210 "num_base_bdevs_operational": 3, 00:11:52.210 "base_bdevs_list": [ 00:11:52.210 { 00:11:52.210 "name": null, 00:11:52.210 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:52.210 "is_configured": false, 00:11:52.210 "data_offset": 0, 00:11:52.210 "data_size": 65536 00:11:52.210 }, 00:11:52.210 { 00:11:52.210 "name": "BaseBdev2", 00:11:52.210 "uuid": "f784b173-69e5-4ed1-b4d1-cc69dcd00ce2", 00:11:52.210 "is_configured": true, 00:11:52.210 "data_offset": 0, 00:11:52.210 "data_size": 65536 00:11:52.210 }, 00:11:52.210 { 00:11:52.210 "name": "BaseBdev3", 00:11:52.210 "uuid": "91c44a5f-7ce8-4207-bd40-3c2891366279", 00:11:52.210 "is_configured": true, 00:11:52.210 "data_offset": 0, 00:11:52.210 "data_size": 65536 00:11:52.210 }, 00:11:52.210 { 00:11:52.210 "name": "BaseBdev4", 00:11:52.210 "uuid": "dfccc5b7-1a80-4e54-8c3b-c466d3013dfd", 00:11:52.210 "is_configured": true, 00:11:52.210 "data_offset": 0, 00:11:52.210 "data_size": 65536 00:11:52.210 } 00:11:52.210 ] 00:11:52.210 }' 00:11:52.210 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:52.210 15:17:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.470 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:11:52.470 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:11:52.470 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:52.470 15:17:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.470 15:17:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.470 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:11:52.470 15:17:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.730 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:11:52.730 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:11:52.730 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:11:52.730 15:17:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.730 15:17:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.730 [2024-11-19 15:17:42.817052] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:11:52.730 15:17:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.730 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:11:52.730 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:11:52.730 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:52.730 15:17:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.730 15:17:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.730 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:11:52.730 15:17:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.730 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:11:52.730 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:11:52.730 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:11:52.730 15:17:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.730 15:17:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.730 [2024-11-19 15:17:42.881582] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:11:52.730 15:17:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.730 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:11:52.730 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:11:52.730 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:52.730 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:11:52.730 15:17:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.730 15:17:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.730 15:17:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.730 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:11:52.730 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:11:52.731 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:11:52.731 15:17:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.731 15:17:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.731 [2024-11-19 15:17:42.957932] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:11:52.731 [2024-11-19 15:17:42.958050] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:52.731 [2024-11-19 15:17:42.979010] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:52.731 [2024-11-19 15:17:42.979117] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:52.731 [2024-11-19 15:17:42.979162] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:11:52.731 15:17:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.731 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:11:52.731 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:11:52.731 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:11:52.731 15:17:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:52.731 15:17:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.731 15:17:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.731 15:17:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.731 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:11:52.731 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:11:52.731 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:11:52.731 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:11:52.731 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:52.731 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:11:52.731 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.731 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.731 BaseBdev2 00:11:52.731 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.731 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:11:52.731 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:11:52.731 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:52.731 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:11:52.731 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:52.731 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:52.731 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:52.731 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.731 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.731 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.731 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:11:52.731 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.731 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.731 [ 00:11:52.731 { 00:11:52.731 "name": "BaseBdev2", 00:11:52.731 "aliases": [ 00:11:52.731 "193db18d-acdf-471c-b8f9-4d3ccb0db4b5" 00:11:52.731 ], 00:11:52.731 "product_name": "Malloc disk", 00:11:52.731 "block_size": 512, 00:11:52.731 "num_blocks": 65536, 00:11:52.731 "uuid": "193db18d-acdf-471c-b8f9-4d3ccb0db4b5", 00:11:52.731 "assigned_rate_limits": { 00:11:52.731 "rw_ios_per_sec": 0, 00:11:52.731 "rw_mbytes_per_sec": 0, 00:11:52.731 "r_mbytes_per_sec": 0, 00:11:52.731 "w_mbytes_per_sec": 0 00:11:52.731 }, 00:11:52.731 "claimed": false, 00:11:52.731 "zoned": false, 00:11:52.731 "supported_io_types": { 00:11:52.731 "read": true, 00:11:52.731 "write": true, 00:11:52.731 "unmap": true, 00:11:52.731 "flush": true, 00:11:52.731 "reset": true, 00:11:52.731 "nvme_admin": false, 00:11:52.731 "nvme_io": false, 00:11:52.731 "nvme_io_md": false, 00:11:52.731 "write_zeroes": true, 00:11:52.731 "zcopy": true, 00:11:52.731 "get_zone_info": false, 00:11:52.992 "zone_management": false, 00:11:52.992 "zone_append": false, 00:11:52.992 "compare": false, 00:11:52.992 "compare_and_write": false, 00:11:52.992 "abort": true, 00:11:52.992 "seek_hole": false, 00:11:52.992 "seek_data": false, 00:11:52.992 "copy": true, 00:11:52.992 "nvme_iov_md": false 00:11:52.992 }, 00:11:52.992 "memory_domains": [ 00:11:52.992 { 00:11:52.992 "dma_device_id": "system", 00:11:52.992 "dma_device_type": 1 00:11:52.992 }, 00:11:52.992 { 00:11:52.992 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:52.992 "dma_device_type": 2 00:11:52.992 } 00:11:52.992 ], 00:11:52.992 "driver_specific": {} 00:11:52.992 } 00:11:52.992 ] 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.992 BaseBdev3 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.992 [ 00:11:52.992 { 00:11:52.992 "name": "BaseBdev3", 00:11:52.992 "aliases": [ 00:11:52.992 "3d665a01-df0f-494d-964e-cb4fdd1c4c0a" 00:11:52.992 ], 00:11:52.992 "product_name": "Malloc disk", 00:11:52.992 "block_size": 512, 00:11:52.992 "num_blocks": 65536, 00:11:52.992 "uuid": "3d665a01-df0f-494d-964e-cb4fdd1c4c0a", 00:11:52.992 "assigned_rate_limits": { 00:11:52.992 "rw_ios_per_sec": 0, 00:11:52.992 "rw_mbytes_per_sec": 0, 00:11:52.992 "r_mbytes_per_sec": 0, 00:11:52.992 "w_mbytes_per_sec": 0 00:11:52.992 }, 00:11:52.992 "claimed": false, 00:11:52.992 "zoned": false, 00:11:52.992 "supported_io_types": { 00:11:52.992 "read": true, 00:11:52.992 "write": true, 00:11:52.992 "unmap": true, 00:11:52.992 "flush": true, 00:11:52.992 "reset": true, 00:11:52.992 "nvme_admin": false, 00:11:52.992 "nvme_io": false, 00:11:52.992 "nvme_io_md": false, 00:11:52.992 "write_zeroes": true, 00:11:52.992 "zcopy": true, 00:11:52.992 "get_zone_info": false, 00:11:52.992 "zone_management": false, 00:11:52.992 "zone_append": false, 00:11:52.992 "compare": false, 00:11:52.992 "compare_and_write": false, 00:11:52.992 "abort": true, 00:11:52.992 "seek_hole": false, 00:11:52.992 "seek_data": false, 00:11:52.992 "copy": true, 00:11:52.992 "nvme_iov_md": false 00:11:52.992 }, 00:11:52.992 "memory_domains": [ 00:11:52.992 { 00:11:52.992 "dma_device_id": "system", 00:11:52.992 "dma_device_type": 1 00:11:52.992 }, 00:11:52.992 { 00:11:52.992 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:52.992 "dma_device_type": 2 00:11:52.992 } 00:11:52.992 ], 00:11:52.992 "driver_specific": {} 00:11:52.992 } 00:11:52.992 ] 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.992 BaseBdev4 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.992 [ 00:11:52.992 { 00:11:52.992 "name": "BaseBdev4", 00:11:52.992 "aliases": [ 00:11:52.992 "da26ef72-6865-41ee-bae1-c6f647d178f5" 00:11:52.992 ], 00:11:52.992 "product_name": "Malloc disk", 00:11:52.992 "block_size": 512, 00:11:52.992 "num_blocks": 65536, 00:11:52.992 "uuid": "da26ef72-6865-41ee-bae1-c6f647d178f5", 00:11:52.992 "assigned_rate_limits": { 00:11:52.992 "rw_ios_per_sec": 0, 00:11:52.992 "rw_mbytes_per_sec": 0, 00:11:52.992 "r_mbytes_per_sec": 0, 00:11:52.992 "w_mbytes_per_sec": 0 00:11:52.992 }, 00:11:52.992 "claimed": false, 00:11:52.992 "zoned": false, 00:11:52.992 "supported_io_types": { 00:11:52.992 "read": true, 00:11:52.992 "write": true, 00:11:52.992 "unmap": true, 00:11:52.992 "flush": true, 00:11:52.992 "reset": true, 00:11:52.992 "nvme_admin": false, 00:11:52.992 "nvme_io": false, 00:11:52.992 "nvme_io_md": false, 00:11:52.992 "write_zeroes": true, 00:11:52.992 "zcopy": true, 00:11:52.992 "get_zone_info": false, 00:11:52.992 "zone_management": false, 00:11:52.992 "zone_append": false, 00:11:52.992 "compare": false, 00:11:52.992 "compare_and_write": false, 00:11:52.992 "abort": true, 00:11:52.992 "seek_hole": false, 00:11:52.992 "seek_data": false, 00:11:52.992 "copy": true, 00:11:52.992 "nvme_iov_md": false 00:11:52.992 }, 00:11:52.992 "memory_domains": [ 00:11:52.992 { 00:11:52.992 "dma_device_id": "system", 00:11:52.992 "dma_device_type": 1 00:11:52.992 }, 00:11:52.992 { 00:11:52.992 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:52.992 "dma_device_type": 2 00:11:52.992 } 00:11:52.992 ], 00:11:52.992 "driver_specific": {} 00:11:52.992 } 00:11:52.992 ] 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:52.992 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.993 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.993 [2024-11-19 15:17:43.206880] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:11:52.993 [2024-11-19 15:17:43.207047] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:11:52.993 [2024-11-19 15:17:43.207106] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:52.993 [2024-11-19 15:17:43.209508] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:52.993 [2024-11-19 15:17:43.209612] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:52.993 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.993 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:52.993 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:52.993 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:52.993 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:52.993 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:52.993 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:52.993 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:52.993 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:52.993 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:52.993 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:52.993 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:52.993 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:52.993 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:52.993 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:52.993 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:52.993 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:52.993 "name": "Existed_Raid", 00:11:52.993 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:52.993 "strip_size_kb": 0, 00:11:52.993 "state": "configuring", 00:11:52.993 "raid_level": "raid1", 00:11:52.993 "superblock": false, 00:11:52.993 "num_base_bdevs": 4, 00:11:52.993 "num_base_bdevs_discovered": 3, 00:11:52.993 "num_base_bdevs_operational": 4, 00:11:52.993 "base_bdevs_list": [ 00:11:52.993 { 00:11:52.993 "name": "BaseBdev1", 00:11:52.993 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:52.993 "is_configured": false, 00:11:52.993 "data_offset": 0, 00:11:52.993 "data_size": 0 00:11:52.993 }, 00:11:52.993 { 00:11:52.993 "name": "BaseBdev2", 00:11:52.993 "uuid": "193db18d-acdf-471c-b8f9-4d3ccb0db4b5", 00:11:52.993 "is_configured": true, 00:11:52.993 "data_offset": 0, 00:11:52.993 "data_size": 65536 00:11:52.993 }, 00:11:52.993 { 00:11:52.993 "name": "BaseBdev3", 00:11:52.993 "uuid": "3d665a01-df0f-494d-964e-cb4fdd1c4c0a", 00:11:52.993 "is_configured": true, 00:11:52.993 "data_offset": 0, 00:11:52.993 "data_size": 65536 00:11:52.993 }, 00:11:52.993 { 00:11:52.993 "name": "BaseBdev4", 00:11:52.993 "uuid": "da26ef72-6865-41ee-bae1-c6f647d178f5", 00:11:52.993 "is_configured": true, 00:11:52.993 "data_offset": 0, 00:11:52.993 "data_size": 65536 00:11:52.993 } 00:11:52.993 ] 00:11:52.993 }' 00:11:52.993 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:52.993 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:53.564 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:11:53.564 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:53.564 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:53.564 [2024-11-19 15:17:43.618137] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:11:53.564 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:53.564 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:53.564 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:53.564 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:53.564 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:53.564 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:53.564 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:53.564 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:53.564 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:53.564 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:53.564 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:53.564 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:53.564 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:53.564 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:53.564 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:53.564 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:53.564 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:53.564 "name": "Existed_Raid", 00:11:53.564 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:53.564 "strip_size_kb": 0, 00:11:53.564 "state": "configuring", 00:11:53.564 "raid_level": "raid1", 00:11:53.564 "superblock": false, 00:11:53.564 "num_base_bdevs": 4, 00:11:53.564 "num_base_bdevs_discovered": 2, 00:11:53.564 "num_base_bdevs_operational": 4, 00:11:53.564 "base_bdevs_list": [ 00:11:53.564 { 00:11:53.564 "name": "BaseBdev1", 00:11:53.564 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:53.564 "is_configured": false, 00:11:53.564 "data_offset": 0, 00:11:53.564 "data_size": 0 00:11:53.564 }, 00:11:53.564 { 00:11:53.564 "name": null, 00:11:53.564 "uuid": "193db18d-acdf-471c-b8f9-4d3ccb0db4b5", 00:11:53.564 "is_configured": false, 00:11:53.564 "data_offset": 0, 00:11:53.564 "data_size": 65536 00:11:53.564 }, 00:11:53.564 { 00:11:53.564 "name": "BaseBdev3", 00:11:53.564 "uuid": "3d665a01-df0f-494d-964e-cb4fdd1c4c0a", 00:11:53.564 "is_configured": true, 00:11:53.564 "data_offset": 0, 00:11:53.564 "data_size": 65536 00:11:53.564 }, 00:11:53.564 { 00:11:53.564 "name": "BaseBdev4", 00:11:53.564 "uuid": "da26ef72-6865-41ee-bae1-c6f647d178f5", 00:11:53.564 "is_configured": true, 00:11:53.564 "data_offset": 0, 00:11:53.564 "data_size": 65536 00:11:53.564 } 00:11:53.564 ] 00:11:53.564 }' 00:11:53.564 15:17:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:53.564 15:17:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:53.825 15:17:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:11:53.825 15:17:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:53.825 15:17:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:53.825 15:17:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:53.825 15:17:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:53.825 15:17:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:11:53.825 15:17:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:11:53.825 15:17:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:53.825 15:17:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:53.825 [2024-11-19 15:17:44.090002] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:53.825 BaseBdev1 00:11:53.825 15:17:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:53.825 15:17:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:11:53.825 15:17:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:11:53.825 15:17:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:53.825 15:17:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:11:53.825 15:17:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:53.825 15:17:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:53.825 15:17:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:53.825 15:17:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:53.825 15:17:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:53.825 15:17:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:53.825 15:17:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:11:53.825 15:17:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:53.825 15:17:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:53.825 [ 00:11:53.825 { 00:11:53.825 "name": "BaseBdev1", 00:11:53.825 "aliases": [ 00:11:53.825 "8fd57454-89cf-4b1d-8779-dc93a254069b" 00:11:53.825 ], 00:11:53.825 "product_name": "Malloc disk", 00:11:53.825 "block_size": 512, 00:11:53.825 "num_blocks": 65536, 00:11:53.825 "uuid": "8fd57454-89cf-4b1d-8779-dc93a254069b", 00:11:53.825 "assigned_rate_limits": { 00:11:53.825 "rw_ios_per_sec": 0, 00:11:53.825 "rw_mbytes_per_sec": 0, 00:11:53.825 "r_mbytes_per_sec": 0, 00:11:53.825 "w_mbytes_per_sec": 0 00:11:53.825 }, 00:11:53.825 "claimed": true, 00:11:53.825 "claim_type": "exclusive_write", 00:11:53.825 "zoned": false, 00:11:53.825 "supported_io_types": { 00:11:53.825 "read": true, 00:11:53.825 "write": true, 00:11:53.825 "unmap": true, 00:11:53.825 "flush": true, 00:11:53.825 "reset": true, 00:11:53.825 "nvme_admin": false, 00:11:53.825 "nvme_io": false, 00:11:53.825 "nvme_io_md": false, 00:11:53.825 "write_zeroes": true, 00:11:53.825 "zcopy": true, 00:11:53.825 "get_zone_info": false, 00:11:53.825 "zone_management": false, 00:11:53.825 "zone_append": false, 00:11:53.825 "compare": false, 00:11:53.825 "compare_and_write": false, 00:11:53.825 "abort": true, 00:11:53.825 "seek_hole": false, 00:11:53.825 "seek_data": false, 00:11:53.825 "copy": true, 00:11:53.825 "nvme_iov_md": false 00:11:53.825 }, 00:11:53.825 "memory_domains": [ 00:11:53.825 { 00:11:53.825 "dma_device_id": "system", 00:11:53.825 "dma_device_type": 1 00:11:53.825 }, 00:11:53.825 { 00:11:53.825 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:53.825 "dma_device_type": 2 00:11:53.825 } 00:11:53.825 ], 00:11:53.825 "driver_specific": {} 00:11:53.825 } 00:11:53.825 ] 00:11:53.825 15:17:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:53.825 15:17:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:11:53.825 15:17:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:53.825 15:17:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:53.825 15:17:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:53.825 15:17:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:53.825 15:17:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:53.825 15:17:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:53.825 15:17:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:53.825 15:17:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:53.825 15:17:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:53.825 15:17:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:53.825 15:17:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:53.825 15:17:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:53.825 15:17:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:53.825 15:17:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:53.825 15:17:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:54.085 15:17:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:54.085 "name": "Existed_Raid", 00:11:54.085 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:54.085 "strip_size_kb": 0, 00:11:54.085 "state": "configuring", 00:11:54.085 "raid_level": "raid1", 00:11:54.085 "superblock": false, 00:11:54.085 "num_base_bdevs": 4, 00:11:54.085 "num_base_bdevs_discovered": 3, 00:11:54.085 "num_base_bdevs_operational": 4, 00:11:54.085 "base_bdevs_list": [ 00:11:54.085 { 00:11:54.085 "name": "BaseBdev1", 00:11:54.085 "uuid": "8fd57454-89cf-4b1d-8779-dc93a254069b", 00:11:54.085 "is_configured": true, 00:11:54.085 "data_offset": 0, 00:11:54.085 "data_size": 65536 00:11:54.085 }, 00:11:54.085 { 00:11:54.085 "name": null, 00:11:54.085 "uuid": "193db18d-acdf-471c-b8f9-4d3ccb0db4b5", 00:11:54.085 "is_configured": false, 00:11:54.085 "data_offset": 0, 00:11:54.085 "data_size": 65536 00:11:54.085 }, 00:11:54.086 { 00:11:54.086 "name": "BaseBdev3", 00:11:54.086 "uuid": "3d665a01-df0f-494d-964e-cb4fdd1c4c0a", 00:11:54.086 "is_configured": true, 00:11:54.086 "data_offset": 0, 00:11:54.086 "data_size": 65536 00:11:54.086 }, 00:11:54.086 { 00:11:54.086 "name": "BaseBdev4", 00:11:54.086 "uuid": "da26ef72-6865-41ee-bae1-c6f647d178f5", 00:11:54.086 "is_configured": true, 00:11:54.086 "data_offset": 0, 00:11:54.086 "data_size": 65536 00:11:54.086 } 00:11:54.086 ] 00:11:54.086 }' 00:11:54.086 15:17:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:54.086 15:17:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:54.346 15:17:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:11:54.346 15:17:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:54.346 15:17:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:54.346 15:17:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:54.346 15:17:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:54.346 15:17:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:11:54.346 15:17:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:11:54.346 15:17:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:54.346 15:17:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:54.346 [2024-11-19 15:17:44.629146] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:11:54.346 15:17:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:54.346 15:17:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:54.346 15:17:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:54.346 15:17:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:54.346 15:17:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:54.346 15:17:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:54.346 15:17:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:54.346 15:17:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:54.346 15:17:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:54.346 15:17:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:54.346 15:17:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:54.346 15:17:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:54.346 15:17:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:54.346 15:17:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:54.346 15:17:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:54.346 15:17:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:54.346 15:17:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:54.346 "name": "Existed_Raid", 00:11:54.346 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:54.346 "strip_size_kb": 0, 00:11:54.346 "state": "configuring", 00:11:54.346 "raid_level": "raid1", 00:11:54.346 "superblock": false, 00:11:54.346 "num_base_bdevs": 4, 00:11:54.346 "num_base_bdevs_discovered": 2, 00:11:54.346 "num_base_bdevs_operational": 4, 00:11:54.346 "base_bdevs_list": [ 00:11:54.346 { 00:11:54.346 "name": "BaseBdev1", 00:11:54.346 "uuid": "8fd57454-89cf-4b1d-8779-dc93a254069b", 00:11:54.346 "is_configured": true, 00:11:54.346 "data_offset": 0, 00:11:54.346 "data_size": 65536 00:11:54.346 }, 00:11:54.346 { 00:11:54.346 "name": null, 00:11:54.346 "uuid": "193db18d-acdf-471c-b8f9-4d3ccb0db4b5", 00:11:54.346 "is_configured": false, 00:11:54.346 "data_offset": 0, 00:11:54.346 "data_size": 65536 00:11:54.346 }, 00:11:54.346 { 00:11:54.346 "name": null, 00:11:54.346 "uuid": "3d665a01-df0f-494d-964e-cb4fdd1c4c0a", 00:11:54.346 "is_configured": false, 00:11:54.346 "data_offset": 0, 00:11:54.346 "data_size": 65536 00:11:54.346 }, 00:11:54.346 { 00:11:54.346 "name": "BaseBdev4", 00:11:54.346 "uuid": "da26ef72-6865-41ee-bae1-c6f647d178f5", 00:11:54.346 "is_configured": true, 00:11:54.346 "data_offset": 0, 00:11:54.346 "data_size": 65536 00:11:54.346 } 00:11:54.346 ] 00:11:54.346 }' 00:11:54.346 15:17:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:54.346 15:17:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:54.916 15:17:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:54.916 15:17:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:11:54.916 15:17:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:54.916 15:17:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:54.916 15:17:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:54.916 15:17:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:11:54.916 15:17:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:11:54.916 15:17:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:54.916 15:17:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:54.916 [2024-11-19 15:17:45.104411] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:54.916 15:17:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:54.916 15:17:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:54.916 15:17:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:54.916 15:17:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:54.916 15:17:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:54.916 15:17:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:54.916 15:17:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:54.916 15:17:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:54.916 15:17:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:54.916 15:17:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:54.916 15:17:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:54.916 15:17:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:54.916 15:17:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:54.916 15:17:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:54.916 15:17:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:54.916 15:17:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:54.916 15:17:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:54.916 "name": "Existed_Raid", 00:11:54.916 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:54.916 "strip_size_kb": 0, 00:11:54.916 "state": "configuring", 00:11:54.916 "raid_level": "raid1", 00:11:54.916 "superblock": false, 00:11:54.916 "num_base_bdevs": 4, 00:11:54.916 "num_base_bdevs_discovered": 3, 00:11:54.916 "num_base_bdevs_operational": 4, 00:11:54.916 "base_bdevs_list": [ 00:11:54.916 { 00:11:54.916 "name": "BaseBdev1", 00:11:54.916 "uuid": "8fd57454-89cf-4b1d-8779-dc93a254069b", 00:11:54.916 "is_configured": true, 00:11:54.916 "data_offset": 0, 00:11:54.916 "data_size": 65536 00:11:54.916 }, 00:11:54.916 { 00:11:54.916 "name": null, 00:11:54.916 "uuid": "193db18d-acdf-471c-b8f9-4d3ccb0db4b5", 00:11:54.916 "is_configured": false, 00:11:54.916 "data_offset": 0, 00:11:54.916 "data_size": 65536 00:11:54.916 }, 00:11:54.916 { 00:11:54.916 "name": "BaseBdev3", 00:11:54.916 "uuid": "3d665a01-df0f-494d-964e-cb4fdd1c4c0a", 00:11:54.916 "is_configured": true, 00:11:54.916 "data_offset": 0, 00:11:54.916 "data_size": 65536 00:11:54.916 }, 00:11:54.916 { 00:11:54.916 "name": "BaseBdev4", 00:11:54.916 "uuid": "da26ef72-6865-41ee-bae1-c6f647d178f5", 00:11:54.916 "is_configured": true, 00:11:54.916 "data_offset": 0, 00:11:54.916 "data_size": 65536 00:11:54.916 } 00:11:54.916 ] 00:11:54.916 }' 00:11:54.916 15:17:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:54.916 15:17:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:55.486 15:17:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:55.486 15:17:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:11:55.486 15:17:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:55.486 15:17:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:55.486 15:17:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:55.486 15:17:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:11:55.486 15:17:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:11:55.486 15:17:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:55.486 15:17:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:55.486 [2024-11-19 15:17:45.599845] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:55.486 15:17:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:55.486 15:17:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:55.486 15:17:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:55.486 15:17:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:55.486 15:17:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:55.486 15:17:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:55.486 15:17:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:55.486 15:17:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:55.486 15:17:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:55.486 15:17:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:55.486 15:17:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:55.486 15:17:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:55.486 15:17:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:55.486 15:17:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:55.486 15:17:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:55.486 15:17:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:55.486 15:17:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:55.486 "name": "Existed_Raid", 00:11:55.486 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:55.487 "strip_size_kb": 0, 00:11:55.487 "state": "configuring", 00:11:55.487 "raid_level": "raid1", 00:11:55.487 "superblock": false, 00:11:55.487 "num_base_bdevs": 4, 00:11:55.487 "num_base_bdevs_discovered": 2, 00:11:55.487 "num_base_bdevs_operational": 4, 00:11:55.487 "base_bdevs_list": [ 00:11:55.487 { 00:11:55.487 "name": null, 00:11:55.487 "uuid": "8fd57454-89cf-4b1d-8779-dc93a254069b", 00:11:55.487 "is_configured": false, 00:11:55.487 "data_offset": 0, 00:11:55.487 "data_size": 65536 00:11:55.487 }, 00:11:55.487 { 00:11:55.487 "name": null, 00:11:55.487 "uuid": "193db18d-acdf-471c-b8f9-4d3ccb0db4b5", 00:11:55.487 "is_configured": false, 00:11:55.487 "data_offset": 0, 00:11:55.487 "data_size": 65536 00:11:55.487 }, 00:11:55.487 { 00:11:55.487 "name": "BaseBdev3", 00:11:55.487 "uuid": "3d665a01-df0f-494d-964e-cb4fdd1c4c0a", 00:11:55.487 "is_configured": true, 00:11:55.487 "data_offset": 0, 00:11:55.487 "data_size": 65536 00:11:55.487 }, 00:11:55.487 { 00:11:55.487 "name": "BaseBdev4", 00:11:55.487 "uuid": "da26ef72-6865-41ee-bae1-c6f647d178f5", 00:11:55.487 "is_configured": true, 00:11:55.487 "data_offset": 0, 00:11:55.487 "data_size": 65536 00:11:55.487 } 00:11:55.487 ] 00:11:55.487 }' 00:11:55.487 15:17:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:55.487 15:17:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:55.747 15:17:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:11:55.747 15:17:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:55.747 15:17:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:55.747 15:17:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:55.747 15:17:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:55.747 15:17:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:11:55.747 15:17:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:11:55.747 15:17:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:55.747 15:17:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:55.747 [2024-11-19 15:17:46.055851] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:55.747 15:17:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:55.747 15:17:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:55.747 15:17:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:55.747 15:17:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:55.747 15:17:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:55.747 15:17:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:55.747 15:17:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:55.747 15:17:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:55.747 15:17:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:55.747 15:17:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:55.747 15:17:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:55.747 15:17:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:55.747 15:17:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:55.747 15:17:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:55.747 15:17:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:56.007 15:17:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:56.007 15:17:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:56.007 "name": "Existed_Raid", 00:11:56.007 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:56.007 "strip_size_kb": 0, 00:11:56.007 "state": "configuring", 00:11:56.007 "raid_level": "raid1", 00:11:56.007 "superblock": false, 00:11:56.007 "num_base_bdevs": 4, 00:11:56.007 "num_base_bdevs_discovered": 3, 00:11:56.007 "num_base_bdevs_operational": 4, 00:11:56.007 "base_bdevs_list": [ 00:11:56.007 { 00:11:56.007 "name": null, 00:11:56.007 "uuid": "8fd57454-89cf-4b1d-8779-dc93a254069b", 00:11:56.007 "is_configured": false, 00:11:56.007 "data_offset": 0, 00:11:56.007 "data_size": 65536 00:11:56.007 }, 00:11:56.007 { 00:11:56.007 "name": "BaseBdev2", 00:11:56.007 "uuid": "193db18d-acdf-471c-b8f9-4d3ccb0db4b5", 00:11:56.007 "is_configured": true, 00:11:56.007 "data_offset": 0, 00:11:56.007 "data_size": 65536 00:11:56.007 }, 00:11:56.007 { 00:11:56.007 "name": "BaseBdev3", 00:11:56.007 "uuid": "3d665a01-df0f-494d-964e-cb4fdd1c4c0a", 00:11:56.007 "is_configured": true, 00:11:56.007 "data_offset": 0, 00:11:56.007 "data_size": 65536 00:11:56.007 }, 00:11:56.007 { 00:11:56.007 "name": "BaseBdev4", 00:11:56.007 "uuid": "da26ef72-6865-41ee-bae1-c6f647d178f5", 00:11:56.007 "is_configured": true, 00:11:56.007 "data_offset": 0, 00:11:56.007 "data_size": 65536 00:11:56.007 } 00:11:56.007 ] 00:11:56.007 }' 00:11:56.007 15:17:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:56.007 15:17:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:56.267 15:17:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:56.268 15:17:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:11:56.268 15:17:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:56.268 15:17:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:56.268 15:17:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:56.268 15:17:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:11:56.268 15:17:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:56.268 15:17:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:11:56.268 15:17:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:56.268 15:17:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:56.268 15:17:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:56.268 15:17:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 8fd57454-89cf-4b1d-8779-dc93a254069b 00:11:56.268 15:17:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:56.268 15:17:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:56.268 [2024-11-19 15:17:46.604134] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:11:56.268 [2024-11-19 15:17:46.604274] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:11:56.268 [2024-11-19 15:17:46.604308] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:11:56.268 [2024-11-19 15:17:46.604635] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:11:56.268 [2024-11-19 15:17:46.604868] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:11:56.268 [2024-11-19 15:17:46.604909] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:11:56.268 [2024-11-19 15:17:46.605177] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:56.268 NewBaseBdev 00:11:56.528 15:17:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:56.528 15:17:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:11:56.528 15:17:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:11:56.528 15:17:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:56.528 15:17:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:11:56.528 15:17:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:56.528 15:17:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:56.528 15:17:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:56.528 15:17:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:56.528 15:17:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:56.528 15:17:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:56.528 15:17:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:11:56.528 15:17:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:56.528 15:17:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:56.528 [ 00:11:56.528 { 00:11:56.528 "name": "NewBaseBdev", 00:11:56.528 "aliases": [ 00:11:56.528 "8fd57454-89cf-4b1d-8779-dc93a254069b" 00:11:56.528 ], 00:11:56.528 "product_name": "Malloc disk", 00:11:56.528 "block_size": 512, 00:11:56.528 "num_blocks": 65536, 00:11:56.528 "uuid": "8fd57454-89cf-4b1d-8779-dc93a254069b", 00:11:56.528 "assigned_rate_limits": { 00:11:56.528 "rw_ios_per_sec": 0, 00:11:56.528 "rw_mbytes_per_sec": 0, 00:11:56.528 "r_mbytes_per_sec": 0, 00:11:56.528 "w_mbytes_per_sec": 0 00:11:56.528 }, 00:11:56.528 "claimed": true, 00:11:56.528 "claim_type": "exclusive_write", 00:11:56.528 "zoned": false, 00:11:56.528 "supported_io_types": { 00:11:56.528 "read": true, 00:11:56.528 "write": true, 00:11:56.528 "unmap": true, 00:11:56.528 "flush": true, 00:11:56.528 "reset": true, 00:11:56.528 "nvme_admin": false, 00:11:56.528 "nvme_io": false, 00:11:56.528 "nvme_io_md": false, 00:11:56.528 "write_zeroes": true, 00:11:56.528 "zcopy": true, 00:11:56.528 "get_zone_info": false, 00:11:56.528 "zone_management": false, 00:11:56.528 "zone_append": false, 00:11:56.528 "compare": false, 00:11:56.528 "compare_and_write": false, 00:11:56.528 "abort": true, 00:11:56.528 "seek_hole": false, 00:11:56.528 "seek_data": false, 00:11:56.528 "copy": true, 00:11:56.528 "nvme_iov_md": false 00:11:56.528 }, 00:11:56.528 "memory_domains": [ 00:11:56.528 { 00:11:56.528 "dma_device_id": "system", 00:11:56.528 "dma_device_type": 1 00:11:56.528 }, 00:11:56.528 { 00:11:56.528 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:56.528 "dma_device_type": 2 00:11:56.528 } 00:11:56.528 ], 00:11:56.528 "driver_specific": {} 00:11:56.528 } 00:11:56.528 ] 00:11:56.529 15:17:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:56.529 15:17:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:11:56.529 15:17:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:11:56.529 15:17:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:56.529 15:17:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:56.529 15:17:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:56.529 15:17:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:56.529 15:17:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:56.529 15:17:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:56.529 15:17:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:56.529 15:17:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:56.529 15:17:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:56.529 15:17:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:56.529 15:17:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:56.529 15:17:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:56.529 15:17:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:56.529 15:17:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:56.529 15:17:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:56.529 "name": "Existed_Raid", 00:11:56.529 "uuid": "9deb9a60-511c-4821-b89b-656a82574b87", 00:11:56.529 "strip_size_kb": 0, 00:11:56.529 "state": "online", 00:11:56.529 "raid_level": "raid1", 00:11:56.529 "superblock": false, 00:11:56.529 "num_base_bdevs": 4, 00:11:56.529 "num_base_bdevs_discovered": 4, 00:11:56.529 "num_base_bdevs_operational": 4, 00:11:56.529 "base_bdevs_list": [ 00:11:56.529 { 00:11:56.529 "name": "NewBaseBdev", 00:11:56.529 "uuid": "8fd57454-89cf-4b1d-8779-dc93a254069b", 00:11:56.529 "is_configured": true, 00:11:56.529 "data_offset": 0, 00:11:56.529 "data_size": 65536 00:11:56.529 }, 00:11:56.529 { 00:11:56.529 "name": "BaseBdev2", 00:11:56.529 "uuid": "193db18d-acdf-471c-b8f9-4d3ccb0db4b5", 00:11:56.529 "is_configured": true, 00:11:56.529 "data_offset": 0, 00:11:56.529 "data_size": 65536 00:11:56.529 }, 00:11:56.529 { 00:11:56.529 "name": "BaseBdev3", 00:11:56.529 "uuid": "3d665a01-df0f-494d-964e-cb4fdd1c4c0a", 00:11:56.529 "is_configured": true, 00:11:56.529 "data_offset": 0, 00:11:56.529 "data_size": 65536 00:11:56.529 }, 00:11:56.529 { 00:11:56.529 "name": "BaseBdev4", 00:11:56.529 "uuid": "da26ef72-6865-41ee-bae1-c6f647d178f5", 00:11:56.529 "is_configured": true, 00:11:56.529 "data_offset": 0, 00:11:56.529 "data_size": 65536 00:11:56.529 } 00:11:56.529 ] 00:11:56.529 }' 00:11:56.529 15:17:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:56.529 15:17:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:56.789 15:17:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:11:56.789 15:17:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:11:56.789 15:17:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:11:56.789 15:17:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:11:56.789 15:17:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:11:56.789 15:17:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:11:56.789 15:17:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:11:56.789 15:17:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:11:56.790 15:17:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:56.790 15:17:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:56.790 [2024-11-19 15:17:47.012235] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:56.790 15:17:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:56.790 15:17:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:11:56.790 "name": "Existed_Raid", 00:11:56.790 "aliases": [ 00:11:56.790 "9deb9a60-511c-4821-b89b-656a82574b87" 00:11:56.790 ], 00:11:56.790 "product_name": "Raid Volume", 00:11:56.790 "block_size": 512, 00:11:56.790 "num_blocks": 65536, 00:11:56.790 "uuid": "9deb9a60-511c-4821-b89b-656a82574b87", 00:11:56.790 "assigned_rate_limits": { 00:11:56.790 "rw_ios_per_sec": 0, 00:11:56.790 "rw_mbytes_per_sec": 0, 00:11:56.790 "r_mbytes_per_sec": 0, 00:11:56.790 "w_mbytes_per_sec": 0 00:11:56.790 }, 00:11:56.790 "claimed": false, 00:11:56.790 "zoned": false, 00:11:56.790 "supported_io_types": { 00:11:56.790 "read": true, 00:11:56.790 "write": true, 00:11:56.790 "unmap": false, 00:11:56.790 "flush": false, 00:11:56.790 "reset": true, 00:11:56.790 "nvme_admin": false, 00:11:56.790 "nvme_io": false, 00:11:56.790 "nvme_io_md": false, 00:11:56.790 "write_zeroes": true, 00:11:56.790 "zcopy": false, 00:11:56.790 "get_zone_info": false, 00:11:56.790 "zone_management": false, 00:11:56.790 "zone_append": false, 00:11:56.790 "compare": false, 00:11:56.790 "compare_and_write": false, 00:11:56.790 "abort": false, 00:11:56.790 "seek_hole": false, 00:11:56.790 "seek_data": false, 00:11:56.790 "copy": false, 00:11:56.790 "nvme_iov_md": false 00:11:56.790 }, 00:11:56.790 "memory_domains": [ 00:11:56.790 { 00:11:56.790 "dma_device_id": "system", 00:11:56.790 "dma_device_type": 1 00:11:56.790 }, 00:11:56.790 { 00:11:56.790 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:56.790 "dma_device_type": 2 00:11:56.790 }, 00:11:56.790 { 00:11:56.790 "dma_device_id": "system", 00:11:56.790 "dma_device_type": 1 00:11:56.790 }, 00:11:56.790 { 00:11:56.790 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:56.790 "dma_device_type": 2 00:11:56.790 }, 00:11:56.790 { 00:11:56.790 "dma_device_id": "system", 00:11:56.790 "dma_device_type": 1 00:11:56.790 }, 00:11:56.790 { 00:11:56.790 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:56.790 "dma_device_type": 2 00:11:56.790 }, 00:11:56.790 { 00:11:56.790 "dma_device_id": "system", 00:11:56.790 "dma_device_type": 1 00:11:56.790 }, 00:11:56.790 { 00:11:56.790 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:56.790 "dma_device_type": 2 00:11:56.790 } 00:11:56.790 ], 00:11:56.790 "driver_specific": { 00:11:56.790 "raid": { 00:11:56.790 "uuid": "9deb9a60-511c-4821-b89b-656a82574b87", 00:11:56.790 "strip_size_kb": 0, 00:11:56.790 "state": "online", 00:11:56.790 "raid_level": "raid1", 00:11:56.790 "superblock": false, 00:11:56.790 "num_base_bdevs": 4, 00:11:56.790 "num_base_bdevs_discovered": 4, 00:11:56.790 "num_base_bdevs_operational": 4, 00:11:56.790 "base_bdevs_list": [ 00:11:56.790 { 00:11:56.790 "name": "NewBaseBdev", 00:11:56.790 "uuid": "8fd57454-89cf-4b1d-8779-dc93a254069b", 00:11:56.790 "is_configured": true, 00:11:56.790 "data_offset": 0, 00:11:56.790 "data_size": 65536 00:11:56.790 }, 00:11:56.790 { 00:11:56.790 "name": "BaseBdev2", 00:11:56.790 "uuid": "193db18d-acdf-471c-b8f9-4d3ccb0db4b5", 00:11:56.790 "is_configured": true, 00:11:56.790 "data_offset": 0, 00:11:56.790 "data_size": 65536 00:11:56.790 }, 00:11:56.790 { 00:11:56.790 "name": "BaseBdev3", 00:11:56.790 "uuid": "3d665a01-df0f-494d-964e-cb4fdd1c4c0a", 00:11:56.790 "is_configured": true, 00:11:56.790 "data_offset": 0, 00:11:56.790 "data_size": 65536 00:11:56.790 }, 00:11:56.790 { 00:11:56.790 "name": "BaseBdev4", 00:11:56.790 "uuid": "da26ef72-6865-41ee-bae1-c6f647d178f5", 00:11:56.790 "is_configured": true, 00:11:56.790 "data_offset": 0, 00:11:56.790 "data_size": 65536 00:11:56.790 } 00:11:56.790 ] 00:11:56.790 } 00:11:56.790 } 00:11:56.790 }' 00:11:56.790 15:17:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:11:56.790 15:17:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:11:56.790 BaseBdev2 00:11:56.790 BaseBdev3 00:11:56.790 BaseBdev4' 00:11:56.790 15:17:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:57.051 [2024-11-19 15:17:47.331305] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:11:57.051 [2024-11-19 15:17:47.331427] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:57.051 [2024-11-19 15:17:47.331545] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:57.051 [2024-11-19 15:17:47.331883] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:57.051 [2024-11-19 15:17:47.331948] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 84011 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 84011 ']' 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 84011 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 84011 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:11:57.051 killing process with pid 84011 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 84011' 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 84011 00:11:57.051 [2024-11-19 15:17:47.380927] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:57.051 15:17:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 84011 00:11:57.311 [2024-11-19 15:17:47.459516] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:57.571 15:17:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:11:57.571 00:11:57.571 real 0m9.604s 00:11:57.571 user 0m16.234s 00:11:57.571 sys 0m1.972s 00:11:57.571 ************************************ 00:11:57.571 END TEST raid_state_function_test 00:11:57.571 ************************************ 00:11:57.571 15:17:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:57.571 15:17:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:57.571 15:17:47 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid1 4 true 00:11:57.571 15:17:47 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:11:57.571 15:17:47 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:57.571 15:17:47 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:57.571 ************************************ 00:11:57.571 START TEST raid_state_function_test_sb 00:11:57.571 ************************************ 00:11:57.571 15:17:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 4 true 00:11:57.571 15:17:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:11:57.571 15:17:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:11:57.571 15:17:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:11:57.571 15:17:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:11:57.571 15:17:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:11:57.571 15:17:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:57.571 15:17:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:11:57.571 15:17:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:11:57.571 15:17:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:57.571 15:17:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:11:57.571 15:17:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:11:57.571 15:17:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:57.571 15:17:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:11:57.571 15:17:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:11:57.571 15:17:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:57.571 15:17:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:11:57.571 15:17:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:11:57.571 15:17:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:57.571 15:17:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:11:57.571 15:17:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:11:57.571 15:17:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:11:57.571 15:17:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:11:57.571 15:17:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:11:57.571 15:17:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:11:57.571 15:17:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:11:57.571 15:17:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:11:57.571 15:17:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:11:57.571 15:17:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:11:57.571 15:17:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:11:57.571 15:17:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=84660 00:11:57.571 15:17:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 84660' 00:11:57.571 Process raid pid: 84660 00:11:57.571 15:17:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 84660 00:11:57.571 15:17:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 84660 ']' 00:11:57.571 15:17:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:57.571 15:17:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:57.571 15:17:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:57.571 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:57.571 15:17:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:57.571 15:17:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:57.831 [2024-11-19 15:17:47.927008] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:11:57.831 [2024-11-19 15:17:47.927197] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:11:57.831 [2024-11-19 15:17:48.064309] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:57.831 [2024-11-19 15:17:48.105732] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:58.101 [2024-11-19 15:17:48.183000] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:58.101 [2024-11-19 15:17:48.183139] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:58.689 15:17:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:58.689 15:17:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:11:58.689 15:17:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:58.689 15:17:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.689 15:17:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:58.689 [2024-11-19 15:17:48.811184] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:11:58.689 [2024-11-19 15:17:48.811378] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:11:58.689 [2024-11-19 15:17:48.811394] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:58.689 [2024-11-19 15:17:48.811405] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:58.689 [2024-11-19 15:17:48.811411] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:11:58.689 [2024-11-19 15:17:48.811424] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:11:58.689 [2024-11-19 15:17:48.811430] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:11:58.689 [2024-11-19 15:17:48.811439] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:11:58.689 15:17:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.689 15:17:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:58.689 15:17:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:58.689 15:17:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:58.689 15:17:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:58.689 15:17:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:58.689 15:17:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:58.689 15:17:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:58.689 15:17:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:58.689 15:17:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:58.689 15:17:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:58.689 15:17:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:58.689 15:17:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.689 15:17:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:58.689 15:17:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:58.689 15:17:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.689 15:17:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:58.689 "name": "Existed_Raid", 00:11:58.689 "uuid": "0e274a9e-4e99-44b8-bb14-a2f3a99fa62b", 00:11:58.689 "strip_size_kb": 0, 00:11:58.689 "state": "configuring", 00:11:58.689 "raid_level": "raid1", 00:11:58.689 "superblock": true, 00:11:58.689 "num_base_bdevs": 4, 00:11:58.689 "num_base_bdevs_discovered": 0, 00:11:58.689 "num_base_bdevs_operational": 4, 00:11:58.689 "base_bdevs_list": [ 00:11:58.689 { 00:11:58.689 "name": "BaseBdev1", 00:11:58.689 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:58.689 "is_configured": false, 00:11:58.689 "data_offset": 0, 00:11:58.689 "data_size": 0 00:11:58.689 }, 00:11:58.689 { 00:11:58.689 "name": "BaseBdev2", 00:11:58.689 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:58.689 "is_configured": false, 00:11:58.689 "data_offset": 0, 00:11:58.689 "data_size": 0 00:11:58.689 }, 00:11:58.689 { 00:11:58.689 "name": "BaseBdev3", 00:11:58.689 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:58.689 "is_configured": false, 00:11:58.689 "data_offset": 0, 00:11:58.689 "data_size": 0 00:11:58.689 }, 00:11:58.689 { 00:11:58.689 "name": "BaseBdev4", 00:11:58.689 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:58.689 "is_configured": false, 00:11:58.689 "data_offset": 0, 00:11:58.689 "data_size": 0 00:11:58.689 } 00:11:58.689 ] 00:11:58.689 }' 00:11:58.689 15:17:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:58.689 15:17:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:58.950 15:17:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:11:58.950 15:17:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.950 15:17:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:58.950 [2024-11-19 15:17:49.246290] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:11:58.950 [2024-11-19 15:17:49.246436] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:11:58.950 15:17:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.950 15:17:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:58.950 15:17:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.950 15:17:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:58.950 [2024-11-19 15:17:49.254248] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:11:58.950 [2024-11-19 15:17:49.254334] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:11:58.950 [2024-11-19 15:17:49.254360] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:58.950 [2024-11-19 15:17:49.254383] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:58.950 [2024-11-19 15:17:49.254403] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:11:58.950 [2024-11-19 15:17:49.254424] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:11:58.950 [2024-11-19 15:17:49.254441] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:11:58.950 [2024-11-19 15:17:49.254463] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:11:58.950 15:17:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.950 15:17:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:11:58.950 15:17:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.950 15:17:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:58.950 [2024-11-19 15:17:49.277087] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:58.950 BaseBdev1 00:11:58.950 15:17:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.950 15:17:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:11:58.950 15:17:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:11:58.950 15:17:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:11:58.950 15:17:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:11:58.950 15:17:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:11:58.950 15:17:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:11:58.950 15:17:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:11:58.950 15:17:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.950 15:17:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:59.210 15:17:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:59.210 15:17:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:11:59.210 15:17:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:59.210 15:17:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:59.210 [ 00:11:59.210 { 00:11:59.210 "name": "BaseBdev1", 00:11:59.210 "aliases": [ 00:11:59.210 "135afbe9-0ff3-4a45-94ab-ac298a5fd90f" 00:11:59.210 ], 00:11:59.210 "product_name": "Malloc disk", 00:11:59.210 "block_size": 512, 00:11:59.210 "num_blocks": 65536, 00:11:59.210 "uuid": "135afbe9-0ff3-4a45-94ab-ac298a5fd90f", 00:11:59.210 "assigned_rate_limits": { 00:11:59.210 "rw_ios_per_sec": 0, 00:11:59.210 "rw_mbytes_per_sec": 0, 00:11:59.210 "r_mbytes_per_sec": 0, 00:11:59.210 "w_mbytes_per_sec": 0 00:11:59.210 }, 00:11:59.210 "claimed": true, 00:11:59.210 "claim_type": "exclusive_write", 00:11:59.210 "zoned": false, 00:11:59.210 "supported_io_types": { 00:11:59.210 "read": true, 00:11:59.210 "write": true, 00:11:59.210 "unmap": true, 00:11:59.210 "flush": true, 00:11:59.210 "reset": true, 00:11:59.210 "nvme_admin": false, 00:11:59.210 "nvme_io": false, 00:11:59.210 "nvme_io_md": false, 00:11:59.210 "write_zeroes": true, 00:11:59.210 "zcopy": true, 00:11:59.210 "get_zone_info": false, 00:11:59.210 "zone_management": false, 00:11:59.210 "zone_append": false, 00:11:59.210 "compare": false, 00:11:59.210 "compare_and_write": false, 00:11:59.210 "abort": true, 00:11:59.210 "seek_hole": false, 00:11:59.210 "seek_data": false, 00:11:59.210 "copy": true, 00:11:59.210 "nvme_iov_md": false 00:11:59.210 }, 00:11:59.210 "memory_domains": [ 00:11:59.210 { 00:11:59.210 "dma_device_id": "system", 00:11:59.210 "dma_device_type": 1 00:11:59.210 }, 00:11:59.210 { 00:11:59.210 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:59.210 "dma_device_type": 2 00:11:59.210 } 00:11:59.210 ], 00:11:59.210 "driver_specific": {} 00:11:59.210 } 00:11:59.210 ] 00:11:59.210 15:17:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:59.210 15:17:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:11:59.210 15:17:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:59.210 15:17:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:59.210 15:17:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:59.210 15:17:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:59.210 15:17:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:59.210 15:17:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:59.210 15:17:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:59.210 15:17:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:59.210 15:17:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:59.210 15:17:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:59.210 15:17:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:59.210 15:17:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:59.210 15:17:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:59.210 15:17:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:59.210 15:17:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:59.210 15:17:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:59.210 "name": "Existed_Raid", 00:11:59.210 "uuid": "7b50a45d-58f9-4c2a-8840-0e38dd9e918b", 00:11:59.210 "strip_size_kb": 0, 00:11:59.210 "state": "configuring", 00:11:59.210 "raid_level": "raid1", 00:11:59.210 "superblock": true, 00:11:59.210 "num_base_bdevs": 4, 00:11:59.210 "num_base_bdevs_discovered": 1, 00:11:59.210 "num_base_bdevs_operational": 4, 00:11:59.210 "base_bdevs_list": [ 00:11:59.210 { 00:11:59.210 "name": "BaseBdev1", 00:11:59.210 "uuid": "135afbe9-0ff3-4a45-94ab-ac298a5fd90f", 00:11:59.210 "is_configured": true, 00:11:59.210 "data_offset": 2048, 00:11:59.210 "data_size": 63488 00:11:59.210 }, 00:11:59.210 { 00:11:59.210 "name": "BaseBdev2", 00:11:59.210 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:59.210 "is_configured": false, 00:11:59.210 "data_offset": 0, 00:11:59.210 "data_size": 0 00:11:59.210 }, 00:11:59.210 { 00:11:59.210 "name": "BaseBdev3", 00:11:59.210 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:59.210 "is_configured": false, 00:11:59.210 "data_offset": 0, 00:11:59.210 "data_size": 0 00:11:59.210 }, 00:11:59.210 { 00:11:59.210 "name": "BaseBdev4", 00:11:59.210 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:59.210 "is_configured": false, 00:11:59.210 "data_offset": 0, 00:11:59.210 "data_size": 0 00:11:59.210 } 00:11:59.210 ] 00:11:59.210 }' 00:11:59.210 15:17:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:59.210 15:17:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:59.471 15:17:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:11:59.471 15:17:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:59.471 15:17:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:59.471 [2024-11-19 15:17:49.748369] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:11:59.471 [2024-11-19 15:17:49.748538] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:11:59.471 15:17:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:59.471 15:17:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:59.471 15:17:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:59.471 15:17:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:59.471 [2024-11-19 15:17:49.760354] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:59.471 [2024-11-19 15:17:49.762577] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:59.471 [2024-11-19 15:17:49.762629] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:59.471 [2024-11-19 15:17:49.762639] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:11:59.471 [2024-11-19 15:17:49.762648] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:11:59.471 [2024-11-19 15:17:49.762654] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:11:59.471 [2024-11-19 15:17:49.762662] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:11:59.471 15:17:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:59.471 15:17:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:11:59.471 15:17:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:59.471 15:17:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:59.471 15:17:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:59.471 15:17:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:59.471 15:17:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:59.471 15:17:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:59.471 15:17:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:59.471 15:17:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:59.471 15:17:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:59.471 15:17:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:59.471 15:17:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:59.471 15:17:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:59.471 15:17:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:59.471 15:17:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:59.471 15:17:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:59.471 15:17:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:59.471 15:17:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:59.471 "name": "Existed_Raid", 00:11:59.471 "uuid": "4852af59-666a-415d-8b4f-4e4087c40cf5", 00:11:59.471 "strip_size_kb": 0, 00:11:59.471 "state": "configuring", 00:11:59.471 "raid_level": "raid1", 00:11:59.471 "superblock": true, 00:11:59.471 "num_base_bdevs": 4, 00:11:59.471 "num_base_bdevs_discovered": 1, 00:11:59.471 "num_base_bdevs_operational": 4, 00:11:59.471 "base_bdevs_list": [ 00:11:59.471 { 00:11:59.471 "name": "BaseBdev1", 00:11:59.471 "uuid": "135afbe9-0ff3-4a45-94ab-ac298a5fd90f", 00:11:59.471 "is_configured": true, 00:11:59.471 "data_offset": 2048, 00:11:59.471 "data_size": 63488 00:11:59.471 }, 00:11:59.471 { 00:11:59.471 "name": "BaseBdev2", 00:11:59.471 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:59.471 "is_configured": false, 00:11:59.471 "data_offset": 0, 00:11:59.471 "data_size": 0 00:11:59.471 }, 00:11:59.471 { 00:11:59.471 "name": "BaseBdev3", 00:11:59.471 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:59.471 "is_configured": false, 00:11:59.471 "data_offset": 0, 00:11:59.471 "data_size": 0 00:11:59.471 }, 00:11:59.472 { 00:11:59.472 "name": "BaseBdev4", 00:11:59.472 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:59.472 "is_configured": false, 00:11:59.472 "data_offset": 0, 00:11:59.472 "data_size": 0 00:11:59.472 } 00:11:59.472 ] 00:11:59.472 }' 00:11:59.472 15:17:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:59.472 15:17:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.042 15:17:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:12:00.042 15:17:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:00.042 15:17:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.042 [2024-11-19 15:17:50.188656] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:00.042 BaseBdev2 00:12:00.042 15:17:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:00.042 15:17:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:12:00.042 15:17:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:12:00.042 15:17:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:12:00.042 15:17:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:12:00.042 15:17:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:12:00.042 15:17:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:12:00.042 15:17:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:12:00.043 15:17:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:00.043 15:17:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.043 15:17:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:00.043 15:17:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:12:00.043 15:17:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:00.043 15:17:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.043 [ 00:12:00.043 { 00:12:00.043 "name": "BaseBdev2", 00:12:00.043 "aliases": [ 00:12:00.043 "94a95976-4149-4c2a-ab5f-e96bbb32fde4" 00:12:00.043 ], 00:12:00.043 "product_name": "Malloc disk", 00:12:00.043 "block_size": 512, 00:12:00.043 "num_blocks": 65536, 00:12:00.043 "uuid": "94a95976-4149-4c2a-ab5f-e96bbb32fde4", 00:12:00.043 "assigned_rate_limits": { 00:12:00.043 "rw_ios_per_sec": 0, 00:12:00.043 "rw_mbytes_per_sec": 0, 00:12:00.043 "r_mbytes_per_sec": 0, 00:12:00.043 "w_mbytes_per_sec": 0 00:12:00.043 }, 00:12:00.043 "claimed": true, 00:12:00.043 "claim_type": "exclusive_write", 00:12:00.043 "zoned": false, 00:12:00.043 "supported_io_types": { 00:12:00.043 "read": true, 00:12:00.043 "write": true, 00:12:00.043 "unmap": true, 00:12:00.043 "flush": true, 00:12:00.043 "reset": true, 00:12:00.043 "nvme_admin": false, 00:12:00.043 "nvme_io": false, 00:12:00.043 "nvme_io_md": false, 00:12:00.043 "write_zeroes": true, 00:12:00.043 "zcopy": true, 00:12:00.043 "get_zone_info": false, 00:12:00.043 "zone_management": false, 00:12:00.043 "zone_append": false, 00:12:00.043 "compare": false, 00:12:00.043 "compare_and_write": false, 00:12:00.043 "abort": true, 00:12:00.043 "seek_hole": false, 00:12:00.043 "seek_data": false, 00:12:00.043 "copy": true, 00:12:00.043 "nvme_iov_md": false 00:12:00.043 }, 00:12:00.043 "memory_domains": [ 00:12:00.043 { 00:12:00.043 "dma_device_id": "system", 00:12:00.043 "dma_device_type": 1 00:12:00.043 }, 00:12:00.043 { 00:12:00.043 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:00.043 "dma_device_type": 2 00:12:00.043 } 00:12:00.043 ], 00:12:00.043 "driver_specific": {} 00:12:00.043 } 00:12:00.043 ] 00:12:00.043 15:17:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:00.043 15:17:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:12:00.043 15:17:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:12:00.043 15:17:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:12:00.043 15:17:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:12:00.043 15:17:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:00.043 15:17:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:00.043 15:17:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:00.043 15:17:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:00.043 15:17:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:00.043 15:17:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:00.043 15:17:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:00.043 15:17:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:00.043 15:17:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:00.043 15:17:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:00.043 15:17:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:00.043 15:17:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:00.043 15:17:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.043 15:17:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:00.043 15:17:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:00.043 "name": "Existed_Raid", 00:12:00.043 "uuid": "4852af59-666a-415d-8b4f-4e4087c40cf5", 00:12:00.043 "strip_size_kb": 0, 00:12:00.043 "state": "configuring", 00:12:00.043 "raid_level": "raid1", 00:12:00.043 "superblock": true, 00:12:00.043 "num_base_bdevs": 4, 00:12:00.043 "num_base_bdevs_discovered": 2, 00:12:00.043 "num_base_bdevs_operational": 4, 00:12:00.043 "base_bdevs_list": [ 00:12:00.043 { 00:12:00.043 "name": "BaseBdev1", 00:12:00.043 "uuid": "135afbe9-0ff3-4a45-94ab-ac298a5fd90f", 00:12:00.043 "is_configured": true, 00:12:00.043 "data_offset": 2048, 00:12:00.043 "data_size": 63488 00:12:00.043 }, 00:12:00.043 { 00:12:00.043 "name": "BaseBdev2", 00:12:00.043 "uuid": "94a95976-4149-4c2a-ab5f-e96bbb32fde4", 00:12:00.043 "is_configured": true, 00:12:00.043 "data_offset": 2048, 00:12:00.043 "data_size": 63488 00:12:00.043 }, 00:12:00.043 { 00:12:00.043 "name": "BaseBdev3", 00:12:00.043 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:00.043 "is_configured": false, 00:12:00.043 "data_offset": 0, 00:12:00.043 "data_size": 0 00:12:00.043 }, 00:12:00.043 { 00:12:00.043 "name": "BaseBdev4", 00:12:00.043 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:00.043 "is_configured": false, 00:12:00.043 "data_offset": 0, 00:12:00.043 "data_size": 0 00:12:00.043 } 00:12:00.043 ] 00:12:00.043 }' 00:12:00.043 15:17:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:00.043 15:17:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.613 15:17:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:12:00.613 15:17:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:00.613 15:17:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.613 [2024-11-19 15:17:50.679960] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:00.613 BaseBdev3 00:12:00.613 15:17:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:00.613 15:17:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:12:00.613 15:17:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:12:00.613 15:17:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:12:00.613 15:17:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:12:00.613 15:17:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:12:00.613 15:17:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:12:00.613 15:17:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:12:00.613 15:17:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:00.613 15:17:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.613 15:17:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:00.613 15:17:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:12:00.613 15:17:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:00.613 15:17:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.613 [ 00:12:00.613 { 00:12:00.613 "name": "BaseBdev3", 00:12:00.613 "aliases": [ 00:12:00.613 "6fe98ded-0e68-4ca7-bf0f-29dd50d1b64f" 00:12:00.613 ], 00:12:00.613 "product_name": "Malloc disk", 00:12:00.613 "block_size": 512, 00:12:00.613 "num_blocks": 65536, 00:12:00.613 "uuid": "6fe98ded-0e68-4ca7-bf0f-29dd50d1b64f", 00:12:00.613 "assigned_rate_limits": { 00:12:00.613 "rw_ios_per_sec": 0, 00:12:00.613 "rw_mbytes_per_sec": 0, 00:12:00.613 "r_mbytes_per_sec": 0, 00:12:00.613 "w_mbytes_per_sec": 0 00:12:00.613 }, 00:12:00.613 "claimed": true, 00:12:00.613 "claim_type": "exclusive_write", 00:12:00.613 "zoned": false, 00:12:00.613 "supported_io_types": { 00:12:00.613 "read": true, 00:12:00.613 "write": true, 00:12:00.613 "unmap": true, 00:12:00.613 "flush": true, 00:12:00.613 "reset": true, 00:12:00.613 "nvme_admin": false, 00:12:00.613 "nvme_io": false, 00:12:00.613 "nvme_io_md": false, 00:12:00.613 "write_zeroes": true, 00:12:00.613 "zcopy": true, 00:12:00.613 "get_zone_info": false, 00:12:00.613 "zone_management": false, 00:12:00.613 "zone_append": false, 00:12:00.613 "compare": false, 00:12:00.613 "compare_and_write": false, 00:12:00.613 "abort": true, 00:12:00.613 "seek_hole": false, 00:12:00.613 "seek_data": false, 00:12:00.613 "copy": true, 00:12:00.613 "nvme_iov_md": false 00:12:00.613 }, 00:12:00.613 "memory_domains": [ 00:12:00.613 { 00:12:00.613 "dma_device_id": "system", 00:12:00.613 "dma_device_type": 1 00:12:00.613 }, 00:12:00.613 { 00:12:00.613 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:00.613 "dma_device_type": 2 00:12:00.613 } 00:12:00.613 ], 00:12:00.613 "driver_specific": {} 00:12:00.613 } 00:12:00.613 ] 00:12:00.613 15:17:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:00.613 15:17:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:12:00.613 15:17:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:12:00.613 15:17:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:12:00.613 15:17:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:12:00.613 15:17:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:00.613 15:17:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:00.613 15:17:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:00.613 15:17:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:00.613 15:17:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:00.613 15:17:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:00.613 15:17:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:00.613 15:17:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:00.613 15:17:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:00.613 15:17:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:00.613 15:17:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:00.613 15:17:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:00.613 15:17:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.613 15:17:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:00.613 15:17:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:00.613 "name": "Existed_Raid", 00:12:00.613 "uuid": "4852af59-666a-415d-8b4f-4e4087c40cf5", 00:12:00.613 "strip_size_kb": 0, 00:12:00.613 "state": "configuring", 00:12:00.613 "raid_level": "raid1", 00:12:00.613 "superblock": true, 00:12:00.613 "num_base_bdevs": 4, 00:12:00.613 "num_base_bdevs_discovered": 3, 00:12:00.613 "num_base_bdevs_operational": 4, 00:12:00.613 "base_bdevs_list": [ 00:12:00.613 { 00:12:00.613 "name": "BaseBdev1", 00:12:00.613 "uuid": "135afbe9-0ff3-4a45-94ab-ac298a5fd90f", 00:12:00.613 "is_configured": true, 00:12:00.613 "data_offset": 2048, 00:12:00.613 "data_size": 63488 00:12:00.613 }, 00:12:00.613 { 00:12:00.613 "name": "BaseBdev2", 00:12:00.613 "uuid": "94a95976-4149-4c2a-ab5f-e96bbb32fde4", 00:12:00.613 "is_configured": true, 00:12:00.613 "data_offset": 2048, 00:12:00.613 "data_size": 63488 00:12:00.613 }, 00:12:00.613 { 00:12:00.613 "name": "BaseBdev3", 00:12:00.613 "uuid": "6fe98ded-0e68-4ca7-bf0f-29dd50d1b64f", 00:12:00.613 "is_configured": true, 00:12:00.613 "data_offset": 2048, 00:12:00.613 "data_size": 63488 00:12:00.613 }, 00:12:00.613 { 00:12:00.613 "name": "BaseBdev4", 00:12:00.613 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:00.613 "is_configured": false, 00:12:00.613 "data_offset": 0, 00:12:00.613 "data_size": 0 00:12:00.613 } 00:12:00.613 ] 00:12:00.613 }' 00:12:00.613 15:17:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:00.613 15:17:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.873 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:12:00.873 15:17:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:00.873 15:17:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.873 [2024-11-19 15:17:51.168301] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:12:00.873 [2024-11-19 15:17:51.168546] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:12:00.873 [2024-11-19 15:17:51.168562] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:00.873 BaseBdev4 00:12:00.873 [2024-11-19 15:17:51.168903] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:12:00.873 [2024-11-19 15:17:51.169081] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:12:00.873 [2024-11-19 15:17:51.169102] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:12:00.873 [2024-11-19 15:17:51.169248] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:00.873 15:17:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:00.873 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:12:00.873 15:17:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:12:00.873 15:17:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:12:00.873 15:17:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:12:00.873 15:17:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:12:00.873 15:17:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:12:00.874 15:17:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:12:00.874 15:17:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:00.874 15:17:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.874 15:17:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:00.874 15:17:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:12:00.874 15:17:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:00.874 15:17:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.874 [ 00:12:00.874 { 00:12:00.874 "name": "BaseBdev4", 00:12:00.874 "aliases": [ 00:12:00.874 "18cdf0c1-e306-4034-9edb-a10f2817321b" 00:12:00.874 ], 00:12:00.874 "product_name": "Malloc disk", 00:12:00.874 "block_size": 512, 00:12:00.874 "num_blocks": 65536, 00:12:00.874 "uuid": "18cdf0c1-e306-4034-9edb-a10f2817321b", 00:12:00.874 "assigned_rate_limits": { 00:12:00.874 "rw_ios_per_sec": 0, 00:12:00.874 "rw_mbytes_per_sec": 0, 00:12:00.874 "r_mbytes_per_sec": 0, 00:12:00.874 "w_mbytes_per_sec": 0 00:12:00.874 }, 00:12:00.874 "claimed": true, 00:12:00.874 "claim_type": "exclusive_write", 00:12:00.874 "zoned": false, 00:12:00.874 "supported_io_types": { 00:12:00.874 "read": true, 00:12:00.874 "write": true, 00:12:00.874 "unmap": true, 00:12:00.874 "flush": true, 00:12:00.874 "reset": true, 00:12:00.874 "nvme_admin": false, 00:12:00.874 "nvme_io": false, 00:12:00.874 "nvme_io_md": false, 00:12:00.874 "write_zeroes": true, 00:12:00.874 "zcopy": true, 00:12:00.874 "get_zone_info": false, 00:12:00.874 "zone_management": false, 00:12:00.874 "zone_append": false, 00:12:00.874 "compare": false, 00:12:00.874 "compare_and_write": false, 00:12:00.874 "abort": true, 00:12:00.874 "seek_hole": false, 00:12:00.874 "seek_data": false, 00:12:00.874 "copy": true, 00:12:00.874 "nvme_iov_md": false 00:12:00.874 }, 00:12:00.874 "memory_domains": [ 00:12:00.874 { 00:12:00.874 "dma_device_id": "system", 00:12:00.874 "dma_device_type": 1 00:12:00.874 }, 00:12:00.874 { 00:12:00.874 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:00.874 "dma_device_type": 2 00:12:00.874 } 00:12:00.874 ], 00:12:00.874 "driver_specific": {} 00:12:00.874 } 00:12:00.874 ] 00:12:00.874 15:17:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:00.874 15:17:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:12:00.874 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:12:00.874 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:12:00.874 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:12:00.874 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:00.874 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:00.874 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:00.874 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:00.874 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:00.874 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:00.874 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:00.874 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:00.874 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:01.134 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:01.134 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:01.134 15:17:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.134 15:17:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:01.134 15:17:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.134 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:01.134 "name": "Existed_Raid", 00:12:01.134 "uuid": "4852af59-666a-415d-8b4f-4e4087c40cf5", 00:12:01.134 "strip_size_kb": 0, 00:12:01.134 "state": "online", 00:12:01.134 "raid_level": "raid1", 00:12:01.134 "superblock": true, 00:12:01.134 "num_base_bdevs": 4, 00:12:01.134 "num_base_bdevs_discovered": 4, 00:12:01.134 "num_base_bdevs_operational": 4, 00:12:01.134 "base_bdevs_list": [ 00:12:01.134 { 00:12:01.134 "name": "BaseBdev1", 00:12:01.134 "uuid": "135afbe9-0ff3-4a45-94ab-ac298a5fd90f", 00:12:01.135 "is_configured": true, 00:12:01.135 "data_offset": 2048, 00:12:01.135 "data_size": 63488 00:12:01.135 }, 00:12:01.135 { 00:12:01.135 "name": "BaseBdev2", 00:12:01.135 "uuid": "94a95976-4149-4c2a-ab5f-e96bbb32fde4", 00:12:01.135 "is_configured": true, 00:12:01.135 "data_offset": 2048, 00:12:01.135 "data_size": 63488 00:12:01.135 }, 00:12:01.135 { 00:12:01.135 "name": "BaseBdev3", 00:12:01.135 "uuid": "6fe98ded-0e68-4ca7-bf0f-29dd50d1b64f", 00:12:01.135 "is_configured": true, 00:12:01.135 "data_offset": 2048, 00:12:01.135 "data_size": 63488 00:12:01.135 }, 00:12:01.135 { 00:12:01.135 "name": "BaseBdev4", 00:12:01.135 "uuid": "18cdf0c1-e306-4034-9edb-a10f2817321b", 00:12:01.135 "is_configured": true, 00:12:01.135 "data_offset": 2048, 00:12:01.135 "data_size": 63488 00:12:01.135 } 00:12:01.135 ] 00:12:01.135 }' 00:12:01.135 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:01.135 15:17:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:01.395 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:12:01.395 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:12:01.395 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:12:01.395 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:12:01.395 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:12:01.395 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:12:01.395 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:12:01.395 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:12:01.395 15:17:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.395 15:17:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:01.395 [2024-11-19 15:17:51.612182] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:01.395 15:17:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.395 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:12:01.395 "name": "Existed_Raid", 00:12:01.395 "aliases": [ 00:12:01.395 "4852af59-666a-415d-8b4f-4e4087c40cf5" 00:12:01.395 ], 00:12:01.395 "product_name": "Raid Volume", 00:12:01.395 "block_size": 512, 00:12:01.395 "num_blocks": 63488, 00:12:01.395 "uuid": "4852af59-666a-415d-8b4f-4e4087c40cf5", 00:12:01.395 "assigned_rate_limits": { 00:12:01.395 "rw_ios_per_sec": 0, 00:12:01.395 "rw_mbytes_per_sec": 0, 00:12:01.395 "r_mbytes_per_sec": 0, 00:12:01.395 "w_mbytes_per_sec": 0 00:12:01.395 }, 00:12:01.395 "claimed": false, 00:12:01.395 "zoned": false, 00:12:01.395 "supported_io_types": { 00:12:01.395 "read": true, 00:12:01.395 "write": true, 00:12:01.395 "unmap": false, 00:12:01.395 "flush": false, 00:12:01.395 "reset": true, 00:12:01.395 "nvme_admin": false, 00:12:01.395 "nvme_io": false, 00:12:01.395 "nvme_io_md": false, 00:12:01.395 "write_zeroes": true, 00:12:01.395 "zcopy": false, 00:12:01.395 "get_zone_info": false, 00:12:01.395 "zone_management": false, 00:12:01.395 "zone_append": false, 00:12:01.395 "compare": false, 00:12:01.395 "compare_and_write": false, 00:12:01.395 "abort": false, 00:12:01.395 "seek_hole": false, 00:12:01.395 "seek_data": false, 00:12:01.395 "copy": false, 00:12:01.395 "nvme_iov_md": false 00:12:01.395 }, 00:12:01.395 "memory_domains": [ 00:12:01.395 { 00:12:01.395 "dma_device_id": "system", 00:12:01.395 "dma_device_type": 1 00:12:01.395 }, 00:12:01.395 { 00:12:01.395 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:01.395 "dma_device_type": 2 00:12:01.395 }, 00:12:01.395 { 00:12:01.395 "dma_device_id": "system", 00:12:01.395 "dma_device_type": 1 00:12:01.395 }, 00:12:01.395 { 00:12:01.395 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:01.395 "dma_device_type": 2 00:12:01.395 }, 00:12:01.395 { 00:12:01.395 "dma_device_id": "system", 00:12:01.395 "dma_device_type": 1 00:12:01.395 }, 00:12:01.395 { 00:12:01.395 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:01.395 "dma_device_type": 2 00:12:01.395 }, 00:12:01.395 { 00:12:01.395 "dma_device_id": "system", 00:12:01.395 "dma_device_type": 1 00:12:01.395 }, 00:12:01.395 { 00:12:01.395 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:01.395 "dma_device_type": 2 00:12:01.395 } 00:12:01.395 ], 00:12:01.395 "driver_specific": { 00:12:01.395 "raid": { 00:12:01.395 "uuid": "4852af59-666a-415d-8b4f-4e4087c40cf5", 00:12:01.395 "strip_size_kb": 0, 00:12:01.395 "state": "online", 00:12:01.395 "raid_level": "raid1", 00:12:01.395 "superblock": true, 00:12:01.395 "num_base_bdevs": 4, 00:12:01.395 "num_base_bdevs_discovered": 4, 00:12:01.395 "num_base_bdevs_operational": 4, 00:12:01.395 "base_bdevs_list": [ 00:12:01.395 { 00:12:01.395 "name": "BaseBdev1", 00:12:01.395 "uuid": "135afbe9-0ff3-4a45-94ab-ac298a5fd90f", 00:12:01.396 "is_configured": true, 00:12:01.396 "data_offset": 2048, 00:12:01.396 "data_size": 63488 00:12:01.396 }, 00:12:01.396 { 00:12:01.396 "name": "BaseBdev2", 00:12:01.396 "uuid": "94a95976-4149-4c2a-ab5f-e96bbb32fde4", 00:12:01.396 "is_configured": true, 00:12:01.396 "data_offset": 2048, 00:12:01.396 "data_size": 63488 00:12:01.396 }, 00:12:01.396 { 00:12:01.396 "name": "BaseBdev3", 00:12:01.396 "uuid": "6fe98ded-0e68-4ca7-bf0f-29dd50d1b64f", 00:12:01.396 "is_configured": true, 00:12:01.396 "data_offset": 2048, 00:12:01.396 "data_size": 63488 00:12:01.396 }, 00:12:01.396 { 00:12:01.396 "name": "BaseBdev4", 00:12:01.396 "uuid": "18cdf0c1-e306-4034-9edb-a10f2817321b", 00:12:01.396 "is_configured": true, 00:12:01.396 "data_offset": 2048, 00:12:01.396 "data_size": 63488 00:12:01.396 } 00:12:01.396 ] 00:12:01.396 } 00:12:01.396 } 00:12:01.396 }' 00:12:01.396 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:12:01.396 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:12:01.396 BaseBdev2 00:12:01.396 BaseBdev3 00:12:01.396 BaseBdev4' 00:12:01.396 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:01.656 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:12:01.656 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:01.656 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:12:01.656 15:17:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.656 15:17:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:01.656 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:01.656 15:17:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.656 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:01.656 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:01.656 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:01.656 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:01.656 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:12:01.656 15:17:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.656 15:17:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:01.656 15:17:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.656 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:01.656 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:01.656 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:01.656 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:12:01.656 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:01.656 15:17:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.656 15:17:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:01.656 15:17:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.656 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:01.656 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:01.656 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:01.656 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:01.656 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:12:01.656 15:17:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.656 15:17:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:01.656 15:17:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.656 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:01.656 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:01.656 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:12:01.656 15:17:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.656 15:17:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:01.656 [2024-11-19 15:17:51.935799] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:12:01.657 15:17:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.657 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:12:01.657 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:12:01.657 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:12:01.657 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:12:01.657 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:12:01.657 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:12:01.657 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:01.657 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:01.657 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:01.657 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:01.657 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:01.657 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:01.657 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:01.657 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:01.657 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:01.657 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:01.657 15:17:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.657 15:17:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:01.657 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:01.657 15:17:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.916 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:01.916 "name": "Existed_Raid", 00:12:01.916 "uuid": "4852af59-666a-415d-8b4f-4e4087c40cf5", 00:12:01.916 "strip_size_kb": 0, 00:12:01.916 "state": "online", 00:12:01.916 "raid_level": "raid1", 00:12:01.916 "superblock": true, 00:12:01.916 "num_base_bdevs": 4, 00:12:01.916 "num_base_bdevs_discovered": 3, 00:12:01.916 "num_base_bdevs_operational": 3, 00:12:01.916 "base_bdevs_list": [ 00:12:01.916 { 00:12:01.916 "name": null, 00:12:01.916 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:01.916 "is_configured": false, 00:12:01.916 "data_offset": 0, 00:12:01.916 "data_size": 63488 00:12:01.916 }, 00:12:01.916 { 00:12:01.916 "name": "BaseBdev2", 00:12:01.916 "uuid": "94a95976-4149-4c2a-ab5f-e96bbb32fde4", 00:12:01.916 "is_configured": true, 00:12:01.916 "data_offset": 2048, 00:12:01.916 "data_size": 63488 00:12:01.916 }, 00:12:01.916 { 00:12:01.916 "name": "BaseBdev3", 00:12:01.916 "uuid": "6fe98ded-0e68-4ca7-bf0f-29dd50d1b64f", 00:12:01.916 "is_configured": true, 00:12:01.916 "data_offset": 2048, 00:12:01.916 "data_size": 63488 00:12:01.916 }, 00:12:01.916 { 00:12:01.916 "name": "BaseBdev4", 00:12:01.916 "uuid": "18cdf0c1-e306-4034-9edb-a10f2817321b", 00:12:01.916 "is_configured": true, 00:12:01.916 "data_offset": 2048, 00:12:01.916 "data_size": 63488 00:12:01.916 } 00:12:01.916 ] 00:12:01.916 }' 00:12:01.916 15:17:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:01.916 15:17:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:02.176 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:12:02.176 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:12:02.176 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:02.176 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:12:02.176 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:02.176 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:02.176 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:02.176 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:12:02.176 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:12:02.176 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:12:02.176 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:02.176 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:02.176 [2024-11-19 15:17:52.451702] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:12:02.176 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:02.176 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:12:02.176 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:12:02.176 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:12:02.176 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:02.176 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:02.176 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:02.176 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:02.176 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:12:02.176 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:12:02.176 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:12:02.176 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:02.176 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:02.436 [2024-11-19 15:17:52.516136] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:12:02.436 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:02.436 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:12:02.436 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:12:02.436 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:02.436 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:12:02.436 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:02.436 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:02.436 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:02.437 [2024-11-19 15:17:52.592908] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:12:02.437 [2024-11-19 15:17:52.593132] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:02.437 [2024-11-19 15:17:52.614627] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:02.437 [2024-11-19 15:17:52.614753] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:02.437 [2024-11-19 15:17:52.614811] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:02.437 BaseBdev2 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:02.437 [ 00:12:02.437 { 00:12:02.437 "name": "BaseBdev2", 00:12:02.437 "aliases": [ 00:12:02.437 "8a2001f0-ea4f-43b2-9b00-2f5e64fe1b60" 00:12:02.437 ], 00:12:02.437 "product_name": "Malloc disk", 00:12:02.437 "block_size": 512, 00:12:02.437 "num_blocks": 65536, 00:12:02.437 "uuid": "8a2001f0-ea4f-43b2-9b00-2f5e64fe1b60", 00:12:02.437 "assigned_rate_limits": { 00:12:02.437 "rw_ios_per_sec": 0, 00:12:02.437 "rw_mbytes_per_sec": 0, 00:12:02.437 "r_mbytes_per_sec": 0, 00:12:02.437 "w_mbytes_per_sec": 0 00:12:02.437 }, 00:12:02.437 "claimed": false, 00:12:02.437 "zoned": false, 00:12:02.437 "supported_io_types": { 00:12:02.437 "read": true, 00:12:02.437 "write": true, 00:12:02.437 "unmap": true, 00:12:02.437 "flush": true, 00:12:02.437 "reset": true, 00:12:02.437 "nvme_admin": false, 00:12:02.437 "nvme_io": false, 00:12:02.437 "nvme_io_md": false, 00:12:02.437 "write_zeroes": true, 00:12:02.437 "zcopy": true, 00:12:02.437 "get_zone_info": false, 00:12:02.437 "zone_management": false, 00:12:02.437 "zone_append": false, 00:12:02.437 "compare": false, 00:12:02.437 "compare_and_write": false, 00:12:02.437 "abort": true, 00:12:02.437 "seek_hole": false, 00:12:02.437 "seek_data": false, 00:12:02.437 "copy": true, 00:12:02.437 "nvme_iov_md": false 00:12:02.437 }, 00:12:02.437 "memory_domains": [ 00:12:02.437 { 00:12:02.437 "dma_device_id": "system", 00:12:02.437 "dma_device_type": 1 00:12:02.437 }, 00:12:02.437 { 00:12:02.437 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:02.437 "dma_device_type": 2 00:12:02.437 } 00:12:02.437 ], 00:12:02.437 "driver_specific": {} 00:12:02.437 } 00:12:02.437 ] 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:02.437 BaseBdev3 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:02.437 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:02.437 [ 00:12:02.437 { 00:12:02.438 "name": "BaseBdev3", 00:12:02.438 "aliases": [ 00:12:02.438 "05e6bdb5-755e-4773-85d7-a9da953f27de" 00:12:02.438 ], 00:12:02.438 "product_name": "Malloc disk", 00:12:02.438 "block_size": 512, 00:12:02.438 "num_blocks": 65536, 00:12:02.438 "uuid": "05e6bdb5-755e-4773-85d7-a9da953f27de", 00:12:02.438 "assigned_rate_limits": { 00:12:02.438 "rw_ios_per_sec": 0, 00:12:02.438 "rw_mbytes_per_sec": 0, 00:12:02.438 "r_mbytes_per_sec": 0, 00:12:02.438 "w_mbytes_per_sec": 0 00:12:02.438 }, 00:12:02.438 "claimed": false, 00:12:02.438 "zoned": false, 00:12:02.438 "supported_io_types": { 00:12:02.438 "read": true, 00:12:02.699 "write": true, 00:12:02.699 "unmap": true, 00:12:02.699 "flush": true, 00:12:02.699 "reset": true, 00:12:02.699 "nvme_admin": false, 00:12:02.699 "nvme_io": false, 00:12:02.699 "nvme_io_md": false, 00:12:02.699 "write_zeroes": true, 00:12:02.699 "zcopy": true, 00:12:02.699 "get_zone_info": false, 00:12:02.699 "zone_management": false, 00:12:02.699 "zone_append": false, 00:12:02.699 "compare": false, 00:12:02.699 "compare_and_write": false, 00:12:02.699 "abort": true, 00:12:02.699 "seek_hole": false, 00:12:02.699 "seek_data": false, 00:12:02.699 "copy": true, 00:12:02.699 "nvme_iov_md": false 00:12:02.699 }, 00:12:02.699 "memory_domains": [ 00:12:02.699 { 00:12:02.699 "dma_device_id": "system", 00:12:02.699 "dma_device_type": 1 00:12:02.699 }, 00:12:02.699 { 00:12:02.699 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:02.699 "dma_device_type": 2 00:12:02.699 } 00:12:02.699 ], 00:12:02.699 "driver_specific": {} 00:12:02.699 } 00:12:02.699 ] 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:02.699 BaseBdev4 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:02.699 [ 00:12:02.699 { 00:12:02.699 "name": "BaseBdev4", 00:12:02.699 "aliases": [ 00:12:02.699 "e5c108b1-ecb2-4a00-99bf-a4b7110f8512" 00:12:02.699 ], 00:12:02.699 "product_name": "Malloc disk", 00:12:02.699 "block_size": 512, 00:12:02.699 "num_blocks": 65536, 00:12:02.699 "uuid": "e5c108b1-ecb2-4a00-99bf-a4b7110f8512", 00:12:02.699 "assigned_rate_limits": { 00:12:02.699 "rw_ios_per_sec": 0, 00:12:02.699 "rw_mbytes_per_sec": 0, 00:12:02.699 "r_mbytes_per_sec": 0, 00:12:02.699 "w_mbytes_per_sec": 0 00:12:02.699 }, 00:12:02.699 "claimed": false, 00:12:02.699 "zoned": false, 00:12:02.699 "supported_io_types": { 00:12:02.699 "read": true, 00:12:02.699 "write": true, 00:12:02.699 "unmap": true, 00:12:02.699 "flush": true, 00:12:02.699 "reset": true, 00:12:02.699 "nvme_admin": false, 00:12:02.699 "nvme_io": false, 00:12:02.699 "nvme_io_md": false, 00:12:02.699 "write_zeroes": true, 00:12:02.699 "zcopy": true, 00:12:02.699 "get_zone_info": false, 00:12:02.699 "zone_management": false, 00:12:02.699 "zone_append": false, 00:12:02.699 "compare": false, 00:12:02.699 "compare_and_write": false, 00:12:02.699 "abort": true, 00:12:02.699 "seek_hole": false, 00:12:02.699 "seek_data": false, 00:12:02.699 "copy": true, 00:12:02.699 "nvme_iov_md": false 00:12:02.699 }, 00:12:02.699 "memory_domains": [ 00:12:02.699 { 00:12:02.699 "dma_device_id": "system", 00:12:02.699 "dma_device_type": 1 00:12:02.699 }, 00:12:02.699 { 00:12:02.699 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:02.699 "dma_device_type": 2 00:12:02.699 } 00:12:02.699 ], 00:12:02.699 "driver_specific": {} 00:12:02.699 } 00:12:02.699 ] 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:02.699 [2024-11-19 15:17:52.856643] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:12:02.699 [2024-11-19 15:17:52.856798] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:12:02.699 [2024-11-19 15:17:52.856858] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:02.699 [2024-11-19 15:17:52.859106] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:02.699 [2024-11-19 15:17:52.859203] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:02.699 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:02.699 "name": "Existed_Raid", 00:12:02.699 "uuid": "c42581f3-7143-4c12-afdd-8d81efeba392", 00:12:02.699 "strip_size_kb": 0, 00:12:02.699 "state": "configuring", 00:12:02.699 "raid_level": "raid1", 00:12:02.699 "superblock": true, 00:12:02.699 "num_base_bdevs": 4, 00:12:02.699 "num_base_bdevs_discovered": 3, 00:12:02.699 "num_base_bdevs_operational": 4, 00:12:02.699 "base_bdevs_list": [ 00:12:02.699 { 00:12:02.699 "name": "BaseBdev1", 00:12:02.699 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:02.699 "is_configured": false, 00:12:02.699 "data_offset": 0, 00:12:02.699 "data_size": 0 00:12:02.699 }, 00:12:02.699 { 00:12:02.699 "name": "BaseBdev2", 00:12:02.699 "uuid": "8a2001f0-ea4f-43b2-9b00-2f5e64fe1b60", 00:12:02.699 "is_configured": true, 00:12:02.699 "data_offset": 2048, 00:12:02.699 "data_size": 63488 00:12:02.699 }, 00:12:02.700 { 00:12:02.700 "name": "BaseBdev3", 00:12:02.700 "uuid": "05e6bdb5-755e-4773-85d7-a9da953f27de", 00:12:02.700 "is_configured": true, 00:12:02.700 "data_offset": 2048, 00:12:02.700 "data_size": 63488 00:12:02.700 }, 00:12:02.700 { 00:12:02.700 "name": "BaseBdev4", 00:12:02.700 "uuid": "e5c108b1-ecb2-4a00-99bf-a4b7110f8512", 00:12:02.700 "is_configured": true, 00:12:02.700 "data_offset": 2048, 00:12:02.700 "data_size": 63488 00:12:02.700 } 00:12:02.700 ] 00:12:02.700 }' 00:12:02.700 15:17:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:02.700 15:17:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:03.268 15:17:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:12:03.268 15:17:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:03.268 15:17:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:03.268 [2024-11-19 15:17:53.319906] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:12:03.268 15:17:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:03.268 15:17:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:12:03.268 15:17:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:03.268 15:17:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:03.268 15:17:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:03.268 15:17:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:03.268 15:17:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:03.268 15:17:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:03.268 15:17:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:03.268 15:17:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:03.268 15:17:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:03.268 15:17:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:03.268 15:17:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:03.268 15:17:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:03.268 15:17:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:03.268 15:17:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:03.268 15:17:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:03.268 "name": "Existed_Raid", 00:12:03.268 "uuid": "c42581f3-7143-4c12-afdd-8d81efeba392", 00:12:03.268 "strip_size_kb": 0, 00:12:03.268 "state": "configuring", 00:12:03.268 "raid_level": "raid1", 00:12:03.268 "superblock": true, 00:12:03.268 "num_base_bdevs": 4, 00:12:03.268 "num_base_bdevs_discovered": 2, 00:12:03.268 "num_base_bdevs_operational": 4, 00:12:03.268 "base_bdevs_list": [ 00:12:03.269 { 00:12:03.269 "name": "BaseBdev1", 00:12:03.269 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:03.269 "is_configured": false, 00:12:03.269 "data_offset": 0, 00:12:03.269 "data_size": 0 00:12:03.269 }, 00:12:03.269 { 00:12:03.269 "name": null, 00:12:03.269 "uuid": "8a2001f0-ea4f-43b2-9b00-2f5e64fe1b60", 00:12:03.269 "is_configured": false, 00:12:03.269 "data_offset": 0, 00:12:03.269 "data_size": 63488 00:12:03.269 }, 00:12:03.269 { 00:12:03.269 "name": "BaseBdev3", 00:12:03.269 "uuid": "05e6bdb5-755e-4773-85d7-a9da953f27de", 00:12:03.269 "is_configured": true, 00:12:03.269 "data_offset": 2048, 00:12:03.269 "data_size": 63488 00:12:03.269 }, 00:12:03.269 { 00:12:03.269 "name": "BaseBdev4", 00:12:03.269 "uuid": "e5c108b1-ecb2-4a00-99bf-a4b7110f8512", 00:12:03.269 "is_configured": true, 00:12:03.269 "data_offset": 2048, 00:12:03.269 "data_size": 63488 00:12:03.269 } 00:12:03.269 ] 00:12:03.269 }' 00:12:03.269 15:17:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:03.269 15:17:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:03.528 15:17:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:12:03.528 15:17:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:03.528 15:17:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:03.528 15:17:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:03.528 15:17:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:03.528 15:17:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:12:03.528 15:17:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:12:03.528 15:17:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:03.528 15:17:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:03.528 [2024-11-19 15:17:53.840054] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:03.528 BaseBdev1 00:12:03.528 15:17:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:03.528 15:17:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:12:03.528 15:17:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:12:03.528 15:17:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:12:03.528 15:17:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:12:03.528 15:17:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:12:03.528 15:17:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:12:03.528 15:17:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:12:03.528 15:17:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:03.528 15:17:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:03.528 15:17:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:03.528 15:17:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:12:03.528 15:17:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:03.528 15:17:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:03.528 [ 00:12:03.528 { 00:12:03.528 "name": "BaseBdev1", 00:12:03.528 "aliases": [ 00:12:03.528 "7a935095-0f87-4a7c-8c5d-08072b399eac" 00:12:03.528 ], 00:12:03.788 "product_name": "Malloc disk", 00:12:03.788 "block_size": 512, 00:12:03.788 "num_blocks": 65536, 00:12:03.788 "uuid": "7a935095-0f87-4a7c-8c5d-08072b399eac", 00:12:03.788 "assigned_rate_limits": { 00:12:03.788 "rw_ios_per_sec": 0, 00:12:03.788 "rw_mbytes_per_sec": 0, 00:12:03.788 "r_mbytes_per_sec": 0, 00:12:03.788 "w_mbytes_per_sec": 0 00:12:03.788 }, 00:12:03.788 "claimed": true, 00:12:03.788 "claim_type": "exclusive_write", 00:12:03.788 "zoned": false, 00:12:03.788 "supported_io_types": { 00:12:03.788 "read": true, 00:12:03.788 "write": true, 00:12:03.788 "unmap": true, 00:12:03.788 "flush": true, 00:12:03.788 "reset": true, 00:12:03.788 "nvme_admin": false, 00:12:03.788 "nvme_io": false, 00:12:03.788 "nvme_io_md": false, 00:12:03.788 "write_zeroes": true, 00:12:03.788 "zcopy": true, 00:12:03.788 "get_zone_info": false, 00:12:03.788 "zone_management": false, 00:12:03.788 "zone_append": false, 00:12:03.788 "compare": false, 00:12:03.788 "compare_and_write": false, 00:12:03.788 "abort": true, 00:12:03.788 "seek_hole": false, 00:12:03.788 "seek_data": false, 00:12:03.788 "copy": true, 00:12:03.788 "nvme_iov_md": false 00:12:03.788 }, 00:12:03.788 "memory_domains": [ 00:12:03.788 { 00:12:03.788 "dma_device_id": "system", 00:12:03.788 "dma_device_type": 1 00:12:03.788 }, 00:12:03.788 { 00:12:03.788 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:03.788 "dma_device_type": 2 00:12:03.788 } 00:12:03.788 ], 00:12:03.788 "driver_specific": {} 00:12:03.788 } 00:12:03.788 ] 00:12:03.788 15:17:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:03.788 15:17:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:12:03.788 15:17:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:12:03.788 15:17:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:03.788 15:17:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:03.788 15:17:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:03.788 15:17:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:03.788 15:17:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:03.788 15:17:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:03.788 15:17:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:03.788 15:17:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:03.788 15:17:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:03.788 15:17:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:03.788 15:17:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:03.788 15:17:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:03.788 15:17:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:03.788 15:17:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:03.788 15:17:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:03.788 "name": "Existed_Raid", 00:12:03.788 "uuid": "c42581f3-7143-4c12-afdd-8d81efeba392", 00:12:03.788 "strip_size_kb": 0, 00:12:03.788 "state": "configuring", 00:12:03.788 "raid_level": "raid1", 00:12:03.788 "superblock": true, 00:12:03.788 "num_base_bdevs": 4, 00:12:03.788 "num_base_bdevs_discovered": 3, 00:12:03.788 "num_base_bdevs_operational": 4, 00:12:03.788 "base_bdevs_list": [ 00:12:03.788 { 00:12:03.788 "name": "BaseBdev1", 00:12:03.788 "uuid": "7a935095-0f87-4a7c-8c5d-08072b399eac", 00:12:03.788 "is_configured": true, 00:12:03.788 "data_offset": 2048, 00:12:03.788 "data_size": 63488 00:12:03.788 }, 00:12:03.788 { 00:12:03.788 "name": null, 00:12:03.788 "uuid": "8a2001f0-ea4f-43b2-9b00-2f5e64fe1b60", 00:12:03.788 "is_configured": false, 00:12:03.788 "data_offset": 0, 00:12:03.788 "data_size": 63488 00:12:03.788 }, 00:12:03.788 { 00:12:03.788 "name": "BaseBdev3", 00:12:03.788 "uuid": "05e6bdb5-755e-4773-85d7-a9da953f27de", 00:12:03.788 "is_configured": true, 00:12:03.789 "data_offset": 2048, 00:12:03.789 "data_size": 63488 00:12:03.789 }, 00:12:03.789 { 00:12:03.789 "name": "BaseBdev4", 00:12:03.789 "uuid": "e5c108b1-ecb2-4a00-99bf-a4b7110f8512", 00:12:03.789 "is_configured": true, 00:12:03.789 "data_offset": 2048, 00:12:03.789 "data_size": 63488 00:12:03.789 } 00:12:03.789 ] 00:12:03.789 }' 00:12:03.789 15:17:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:03.789 15:17:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:04.048 15:17:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:04.048 15:17:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:12:04.048 15:17:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:04.048 15:17:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:04.048 15:17:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:04.048 15:17:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:12:04.049 15:17:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:12:04.049 15:17:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:04.049 15:17:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:04.049 [2024-11-19 15:17:54.363392] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:12:04.049 15:17:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:04.049 15:17:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:12:04.049 15:17:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:04.049 15:17:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:04.049 15:17:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:04.049 15:17:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:04.049 15:17:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:04.049 15:17:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:04.049 15:17:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:04.049 15:17:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:04.049 15:17:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:04.049 15:17:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:04.049 15:17:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:04.049 15:17:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:04.049 15:17:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:04.307 15:17:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:04.307 15:17:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:04.307 "name": "Existed_Raid", 00:12:04.307 "uuid": "c42581f3-7143-4c12-afdd-8d81efeba392", 00:12:04.307 "strip_size_kb": 0, 00:12:04.307 "state": "configuring", 00:12:04.307 "raid_level": "raid1", 00:12:04.307 "superblock": true, 00:12:04.307 "num_base_bdevs": 4, 00:12:04.307 "num_base_bdevs_discovered": 2, 00:12:04.307 "num_base_bdevs_operational": 4, 00:12:04.307 "base_bdevs_list": [ 00:12:04.307 { 00:12:04.307 "name": "BaseBdev1", 00:12:04.307 "uuid": "7a935095-0f87-4a7c-8c5d-08072b399eac", 00:12:04.307 "is_configured": true, 00:12:04.307 "data_offset": 2048, 00:12:04.307 "data_size": 63488 00:12:04.307 }, 00:12:04.307 { 00:12:04.307 "name": null, 00:12:04.307 "uuid": "8a2001f0-ea4f-43b2-9b00-2f5e64fe1b60", 00:12:04.307 "is_configured": false, 00:12:04.307 "data_offset": 0, 00:12:04.307 "data_size": 63488 00:12:04.307 }, 00:12:04.307 { 00:12:04.307 "name": null, 00:12:04.307 "uuid": "05e6bdb5-755e-4773-85d7-a9da953f27de", 00:12:04.307 "is_configured": false, 00:12:04.307 "data_offset": 0, 00:12:04.307 "data_size": 63488 00:12:04.307 }, 00:12:04.307 { 00:12:04.307 "name": "BaseBdev4", 00:12:04.307 "uuid": "e5c108b1-ecb2-4a00-99bf-a4b7110f8512", 00:12:04.307 "is_configured": true, 00:12:04.307 "data_offset": 2048, 00:12:04.307 "data_size": 63488 00:12:04.307 } 00:12:04.307 ] 00:12:04.307 }' 00:12:04.307 15:17:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:04.307 15:17:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:04.567 15:17:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:04.567 15:17:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:04.567 15:17:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:12:04.567 15:17:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:04.567 15:17:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:04.567 15:17:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:12:04.567 15:17:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:12:04.567 15:17:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:04.567 15:17:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:04.567 [2024-11-19 15:17:54.858550] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:04.567 15:17:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:04.567 15:17:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:12:04.567 15:17:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:04.567 15:17:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:04.567 15:17:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:04.567 15:17:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:04.567 15:17:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:04.567 15:17:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:04.567 15:17:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:04.567 15:17:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:04.567 15:17:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:04.567 15:17:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:04.567 15:17:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:04.567 15:17:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:04.567 15:17:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:04.567 15:17:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:04.826 15:17:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:04.826 "name": "Existed_Raid", 00:12:04.826 "uuid": "c42581f3-7143-4c12-afdd-8d81efeba392", 00:12:04.826 "strip_size_kb": 0, 00:12:04.826 "state": "configuring", 00:12:04.826 "raid_level": "raid1", 00:12:04.826 "superblock": true, 00:12:04.826 "num_base_bdevs": 4, 00:12:04.826 "num_base_bdevs_discovered": 3, 00:12:04.826 "num_base_bdevs_operational": 4, 00:12:04.826 "base_bdevs_list": [ 00:12:04.826 { 00:12:04.826 "name": "BaseBdev1", 00:12:04.826 "uuid": "7a935095-0f87-4a7c-8c5d-08072b399eac", 00:12:04.826 "is_configured": true, 00:12:04.826 "data_offset": 2048, 00:12:04.826 "data_size": 63488 00:12:04.826 }, 00:12:04.826 { 00:12:04.826 "name": null, 00:12:04.826 "uuid": "8a2001f0-ea4f-43b2-9b00-2f5e64fe1b60", 00:12:04.826 "is_configured": false, 00:12:04.826 "data_offset": 0, 00:12:04.826 "data_size": 63488 00:12:04.826 }, 00:12:04.826 { 00:12:04.826 "name": "BaseBdev3", 00:12:04.826 "uuid": "05e6bdb5-755e-4773-85d7-a9da953f27de", 00:12:04.826 "is_configured": true, 00:12:04.826 "data_offset": 2048, 00:12:04.826 "data_size": 63488 00:12:04.826 }, 00:12:04.826 { 00:12:04.827 "name": "BaseBdev4", 00:12:04.827 "uuid": "e5c108b1-ecb2-4a00-99bf-a4b7110f8512", 00:12:04.827 "is_configured": true, 00:12:04.827 "data_offset": 2048, 00:12:04.827 "data_size": 63488 00:12:04.827 } 00:12:04.827 ] 00:12:04.827 }' 00:12:04.827 15:17:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:04.827 15:17:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:05.086 15:17:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:12:05.086 15:17:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:05.086 15:17:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:05.086 15:17:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:05.086 15:17:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:05.086 15:17:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:12:05.086 15:17:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:12:05.086 15:17:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:05.086 15:17:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:05.086 [2024-11-19 15:17:55.353781] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:12:05.086 15:17:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:05.086 15:17:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:12:05.086 15:17:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:05.086 15:17:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:05.086 15:17:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:05.086 15:17:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:05.086 15:17:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:05.086 15:17:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:05.086 15:17:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:05.086 15:17:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:05.086 15:17:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:05.086 15:17:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:05.086 15:17:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:05.086 15:17:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:05.086 15:17:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:05.086 15:17:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:05.346 15:17:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:05.346 "name": "Existed_Raid", 00:12:05.346 "uuid": "c42581f3-7143-4c12-afdd-8d81efeba392", 00:12:05.346 "strip_size_kb": 0, 00:12:05.346 "state": "configuring", 00:12:05.346 "raid_level": "raid1", 00:12:05.346 "superblock": true, 00:12:05.346 "num_base_bdevs": 4, 00:12:05.346 "num_base_bdevs_discovered": 2, 00:12:05.346 "num_base_bdevs_operational": 4, 00:12:05.346 "base_bdevs_list": [ 00:12:05.346 { 00:12:05.346 "name": null, 00:12:05.346 "uuid": "7a935095-0f87-4a7c-8c5d-08072b399eac", 00:12:05.346 "is_configured": false, 00:12:05.346 "data_offset": 0, 00:12:05.346 "data_size": 63488 00:12:05.346 }, 00:12:05.346 { 00:12:05.346 "name": null, 00:12:05.346 "uuid": "8a2001f0-ea4f-43b2-9b00-2f5e64fe1b60", 00:12:05.346 "is_configured": false, 00:12:05.346 "data_offset": 0, 00:12:05.346 "data_size": 63488 00:12:05.346 }, 00:12:05.346 { 00:12:05.346 "name": "BaseBdev3", 00:12:05.346 "uuid": "05e6bdb5-755e-4773-85d7-a9da953f27de", 00:12:05.346 "is_configured": true, 00:12:05.346 "data_offset": 2048, 00:12:05.346 "data_size": 63488 00:12:05.346 }, 00:12:05.346 { 00:12:05.346 "name": "BaseBdev4", 00:12:05.346 "uuid": "e5c108b1-ecb2-4a00-99bf-a4b7110f8512", 00:12:05.346 "is_configured": true, 00:12:05.346 "data_offset": 2048, 00:12:05.346 "data_size": 63488 00:12:05.346 } 00:12:05.346 ] 00:12:05.346 }' 00:12:05.346 15:17:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:05.346 15:17:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:05.606 15:17:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:05.606 15:17:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:12:05.606 15:17:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:05.606 15:17:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:05.606 15:17:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:05.606 15:17:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:12:05.606 15:17:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:12:05.606 15:17:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:05.606 15:17:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:05.606 [2024-11-19 15:17:55.852564] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:05.606 15:17:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:05.606 15:17:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:12:05.606 15:17:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:05.606 15:17:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:05.606 15:17:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:05.606 15:17:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:05.606 15:17:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:05.606 15:17:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:05.606 15:17:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:05.606 15:17:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:05.606 15:17:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:05.606 15:17:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:05.606 15:17:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:05.606 15:17:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:05.606 15:17:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:05.606 15:17:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:05.606 15:17:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:05.606 "name": "Existed_Raid", 00:12:05.606 "uuid": "c42581f3-7143-4c12-afdd-8d81efeba392", 00:12:05.606 "strip_size_kb": 0, 00:12:05.606 "state": "configuring", 00:12:05.606 "raid_level": "raid1", 00:12:05.606 "superblock": true, 00:12:05.606 "num_base_bdevs": 4, 00:12:05.606 "num_base_bdevs_discovered": 3, 00:12:05.606 "num_base_bdevs_operational": 4, 00:12:05.606 "base_bdevs_list": [ 00:12:05.606 { 00:12:05.606 "name": null, 00:12:05.606 "uuid": "7a935095-0f87-4a7c-8c5d-08072b399eac", 00:12:05.606 "is_configured": false, 00:12:05.606 "data_offset": 0, 00:12:05.606 "data_size": 63488 00:12:05.606 }, 00:12:05.606 { 00:12:05.606 "name": "BaseBdev2", 00:12:05.606 "uuid": "8a2001f0-ea4f-43b2-9b00-2f5e64fe1b60", 00:12:05.606 "is_configured": true, 00:12:05.606 "data_offset": 2048, 00:12:05.606 "data_size": 63488 00:12:05.606 }, 00:12:05.606 { 00:12:05.606 "name": "BaseBdev3", 00:12:05.606 "uuid": "05e6bdb5-755e-4773-85d7-a9da953f27de", 00:12:05.606 "is_configured": true, 00:12:05.606 "data_offset": 2048, 00:12:05.606 "data_size": 63488 00:12:05.606 }, 00:12:05.606 { 00:12:05.606 "name": "BaseBdev4", 00:12:05.606 "uuid": "e5c108b1-ecb2-4a00-99bf-a4b7110f8512", 00:12:05.606 "is_configured": true, 00:12:05.606 "data_offset": 2048, 00:12:05.606 "data_size": 63488 00:12:05.606 } 00:12:05.606 ] 00:12:05.606 }' 00:12:05.606 15:17:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:05.606 15:17:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:06.175 15:17:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:06.175 15:17:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:06.175 15:17:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:06.175 15:17:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:12:06.175 15:17:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:06.175 15:17:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:12:06.175 15:17:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:06.175 15:17:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:12:06.175 15:17:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:06.175 15:17:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:06.175 15:17:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:06.175 15:17:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 7a935095-0f87-4a7c-8c5d-08072b399eac 00:12:06.175 15:17:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:06.175 15:17:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:06.175 [2024-11-19 15:17:56.380662] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:12:06.175 [2024-11-19 15:17:56.380871] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:12:06.175 [2024-11-19 15:17:56.380890] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:06.175 [2024-11-19 15:17:56.381188] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:12:06.175 NewBaseBdev 00:12:06.175 [2024-11-19 15:17:56.381333] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:12:06.175 [2024-11-19 15:17:56.381343] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:12:06.175 [2024-11-19 15:17:56.381452] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:06.175 15:17:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:06.175 15:17:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:12:06.175 15:17:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:12:06.175 15:17:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:12:06.175 15:17:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:12:06.176 15:17:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:12:06.176 15:17:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:12:06.176 15:17:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:12:06.176 15:17:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:06.176 15:17:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:06.176 15:17:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:06.176 15:17:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:12:06.176 15:17:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:06.176 15:17:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:06.176 [ 00:12:06.176 { 00:12:06.176 "name": "NewBaseBdev", 00:12:06.176 "aliases": [ 00:12:06.176 "7a935095-0f87-4a7c-8c5d-08072b399eac" 00:12:06.176 ], 00:12:06.176 "product_name": "Malloc disk", 00:12:06.176 "block_size": 512, 00:12:06.176 "num_blocks": 65536, 00:12:06.176 "uuid": "7a935095-0f87-4a7c-8c5d-08072b399eac", 00:12:06.176 "assigned_rate_limits": { 00:12:06.176 "rw_ios_per_sec": 0, 00:12:06.176 "rw_mbytes_per_sec": 0, 00:12:06.176 "r_mbytes_per_sec": 0, 00:12:06.176 "w_mbytes_per_sec": 0 00:12:06.176 }, 00:12:06.176 "claimed": true, 00:12:06.176 "claim_type": "exclusive_write", 00:12:06.176 "zoned": false, 00:12:06.176 "supported_io_types": { 00:12:06.176 "read": true, 00:12:06.176 "write": true, 00:12:06.176 "unmap": true, 00:12:06.176 "flush": true, 00:12:06.176 "reset": true, 00:12:06.176 "nvme_admin": false, 00:12:06.176 "nvme_io": false, 00:12:06.176 "nvme_io_md": false, 00:12:06.176 "write_zeroes": true, 00:12:06.176 "zcopy": true, 00:12:06.176 "get_zone_info": false, 00:12:06.176 "zone_management": false, 00:12:06.176 "zone_append": false, 00:12:06.176 "compare": false, 00:12:06.176 "compare_and_write": false, 00:12:06.176 "abort": true, 00:12:06.176 "seek_hole": false, 00:12:06.176 "seek_data": false, 00:12:06.176 "copy": true, 00:12:06.176 "nvme_iov_md": false 00:12:06.176 }, 00:12:06.176 "memory_domains": [ 00:12:06.176 { 00:12:06.176 "dma_device_id": "system", 00:12:06.176 "dma_device_type": 1 00:12:06.176 }, 00:12:06.176 { 00:12:06.176 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:06.176 "dma_device_type": 2 00:12:06.176 } 00:12:06.176 ], 00:12:06.176 "driver_specific": {} 00:12:06.176 } 00:12:06.176 ] 00:12:06.176 15:17:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:06.176 15:17:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:12:06.176 15:17:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:12:06.176 15:17:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:12:06.176 15:17:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:06.176 15:17:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:06.176 15:17:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:06.176 15:17:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:06.176 15:17:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:06.176 15:17:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:06.176 15:17:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:06.176 15:17:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:06.176 15:17:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:12:06.176 15:17:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:06.176 15:17:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:06.176 15:17:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:06.176 15:17:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:06.176 15:17:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:06.176 "name": "Existed_Raid", 00:12:06.176 "uuid": "c42581f3-7143-4c12-afdd-8d81efeba392", 00:12:06.176 "strip_size_kb": 0, 00:12:06.176 "state": "online", 00:12:06.176 "raid_level": "raid1", 00:12:06.176 "superblock": true, 00:12:06.176 "num_base_bdevs": 4, 00:12:06.176 "num_base_bdevs_discovered": 4, 00:12:06.176 "num_base_bdevs_operational": 4, 00:12:06.176 "base_bdevs_list": [ 00:12:06.176 { 00:12:06.176 "name": "NewBaseBdev", 00:12:06.176 "uuid": "7a935095-0f87-4a7c-8c5d-08072b399eac", 00:12:06.176 "is_configured": true, 00:12:06.176 "data_offset": 2048, 00:12:06.176 "data_size": 63488 00:12:06.176 }, 00:12:06.176 { 00:12:06.176 "name": "BaseBdev2", 00:12:06.176 "uuid": "8a2001f0-ea4f-43b2-9b00-2f5e64fe1b60", 00:12:06.176 "is_configured": true, 00:12:06.176 "data_offset": 2048, 00:12:06.176 "data_size": 63488 00:12:06.176 }, 00:12:06.176 { 00:12:06.176 "name": "BaseBdev3", 00:12:06.176 "uuid": "05e6bdb5-755e-4773-85d7-a9da953f27de", 00:12:06.176 "is_configured": true, 00:12:06.176 "data_offset": 2048, 00:12:06.176 "data_size": 63488 00:12:06.176 }, 00:12:06.176 { 00:12:06.176 "name": "BaseBdev4", 00:12:06.176 "uuid": "e5c108b1-ecb2-4a00-99bf-a4b7110f8512", 00:12:06.176 "is_configured": true, 00:12:06.176 "data_offset": 2048, 00:12:06.176 "data_size": 63488 00:12:06.176 } 00:12:06.176 ] 00:12:06.176 }' 00:12:06.176 15:17:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:06.176 15:17:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:06.748 15:17:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:12:06.748 15:17:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:12:06.748 15:17:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:12:06.748 15:17:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:12:06.748 15:17:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:12:06.748 15:17:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:12:06.748 15:17:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:12:06.748 15:17:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:12:06.748 15:17:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:06.748 15:17:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:06.748 [2024-11-19 15:17:56.848320] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:06.748 15:17:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:06.748 15:17:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:12:06.748 "name": "Existed_Raid", 00:12:06.748 "aliases": [ 00:12:06.748 "c42581f3-7143-4c12-afdd-8d81efeba392" 00:12:06.748 ], 00:12:06.748 "product_name": "Raid Volume", 00:12:06.748 "block_size": 512, 00:12:06.748 "num_blocks": 63488, 00:12:06.748 "uuid": "c42581f3-7143-4c12-afdd-8d81efeba392", 00:12:06.748 "assigned_rate_limits": { 00:12:06.748 "rw_ios_per_sec": 0, 00:12:06.748 "rw_mbytes_per_sec": 0, 00:12:06.748 "r_mbytes_per_sec": 0, 00:12:06.748 "w_mbytes_per_sec": 0 00:12:06.748 }, 00:12:06.748 "claimed": false, 00:12:06.748 "zoned": false, 00:12:06.748 "supported_io_types": { 00:12:06.748 "read": true, 00:12:06.748 "write": true, 00:12:06.748 "unmap": false, 00:12:06.748 "flush": false, 00:12:06.748 "reset": true, 00:12:06.748 "nvme_admin": false, 00:12:06.748 "nvme_io": false, 00:12:06.748 "nvme_io_md": false, 00:12:06.748 "write_zeroes": true, 00:12:06.748 "zcopy": false, 00:12:06.748 "get_zone_info": false, 00:12:06.748 "zone_management": false, 00:12:06.748 "zone_append": false, 00:12:06.748 "compare": false, 00:12:06.748 "compare_and_write": false, 00:12:06.748 "abort": false, 00:12:06.748 "seek_hole": false, 00:12:06.748 "seek_data": false, 00:12:06.748 "copy": false, 00:12:06.748 "nvme_iov_md": false 00:12:06.748 }, 00:12:06.748 "memory_domains": [ 00:12:06.748 { 00:12:06.748 "dma_device_id": "system", 00:12:06.748 "dma_device_type": 1 00:12:06.748 }, 00:12:06.748 { 00:12:06.748 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:06.748 "dma_device_type": 2 00:12:06.748 }, 00:12:06.748 { 00:12:06.748 "dma_device_id": "system", 00:12:06.748 "dma_device_type": 1 00:12:06.748 }, 00:12:06.748 { 00:12:06.748 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:06.748 "dma_device_type": 2 00:12:06.748 }, 00:12:06.748 { 00:12:06.748 "dma_device_id": "system", 00:12:06.748 "dma_device_type": 1 00:12:06.748 }, 00:12:06.748 { 00:12:06.748 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:06.748 "dma_device_type": 2 00:12:06.748 }, 00:12:06.748 { 00:12:06.748 "dma_device_id": "system", 00:12:06.748 "dma_device_type": 1 00:12:06.748 }, 00:12:06.748 { 00:12:06.748 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:06.748 "dma_device_type": 2 00:12:06.748 } 00:12:06.748 ], 00:12:06.748 "driver_specific": { 00:12:06.748 "raid": { 00:12:06.748 "uuid": "c42581f3-7143-4c12-afdd-8d81efeba392", 00:12:06.748 "strip_size_kb": 0, 00:12:06.748 "state": "online", 00:12:06.748 "raid_level": "raid1", 00:12:06.748 "superblock": true, 00:12:06.748 "num_base_bdevs": 4, 00:12:06.748 "num_base_bdevs_discovered": 4, 00:12:06.748 "num_base_bdevs_operational": 4, 00:12:06.748 "base_bdevs_list": [ 00:12:06.748 { 00:12:06.748 "name": "NewBaseBdev", 00:12:06.748 "uuid": "7a935095-0f87-4a7c-8c5d-08072b399eac", 00:12:06.748 "is_configured": true, 00:12:06.748 "data_offset": 2048, 00:12:06.748 "data_size": 63488 00:12:06.748 }, 00:12:06.748 { 00:12:06.748 "name": "BaseBdev2", 00:12:06.748 "uuid": "8a2001f0-ea4f-43b2-9b00-2f5e64fe1b60", 00:12:06.748 "is_configured": true, 00:12:06.748 "data_offset": 2048, 00:12:06.748 "data_size": 63488 00:12:06.748 }, 00:12:06.748 { 00:12:06.748 "name": "BaseBdev3", 00:12:06.748 "uuid": "05e6bdb5-755e-4773-85d7-a9da953f27de", 00:12:06.748 "is_configured": true, 00:12:06.748 "data_offset": 2048, 00:12:06.748 "data_size": 63488 00:12:06.748 }, 00:12:06.748 { 00:12:06.748 "name": "BaseBdev4", 00:12:06.748 "uuid": "e5c108b1-ecb2-4a00-99bf-a4b7110f8512", 00:12:06.748 "is_configured": true, 00:12:06.748 "data_offset": 2048, 00:12:06.748 "data_size": 63488 00:12:06.748 } 00:12:06.748 ] 00:12:06.748 } 00:12:06.748 } 00:12:06.748 }' 00:12:06.748 15:17:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:12:06.748 15:17:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:12:06.748 BaseBdev2 00:12:06.748 BaseBdev3 00:12:06.748 BaseBdev4' 00:12:06.748 15:17:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:06.748 15:17:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:12:06.748 15:17:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:06.748 15:17:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:12:06.748 15:17:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:06.748 15:17:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:06.748 15:17:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:06.748 15:17:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:06.748 15:17:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:06.748 15:17:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:06.748 15:17:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:06.748 15:17:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:12:06.748 15:17:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:06.748 15:17:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:06.748 15:17:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:06.749 15:17:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:06.749 15:17:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:06.749 15:17:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:06.749 15:17:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:06.749 15:17:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:06.749 15:17:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:12:06.749 15:17:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:06.749 15:17:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:07.008 15:17:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:07.008 15:17:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:07.008 15:17:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:07.008 15:17:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:07.008 15:17:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:07.008 15:17:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:12:07.008 15:17:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:07.008 15:17:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:07.008 15:17:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:07.008 15:17:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:07.008 15:17:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:07.008 15:17:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:12:07.008 15:17:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:07.008 15:17:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:07.008 [2024-11-19 15:17:57.151647] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:12:07.008 [2024-11-19 15:17:57.151804] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:07.008 [2024-11-19 15:17:57.151925] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:07.008 [2024-11-19 15:17:57.152256] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:07.008 [2024-11-19 15:17:57.152279] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:12:07.008 15:17:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:07.008 15:17:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 84660 00:12:07.008 15:17:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 84660 ']' 00:12:07.008 15:17:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 84660 00:12:07.008 15:17:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:12:07.008 15:17:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:12:07.008 15:17:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 84660 00:12:07.008 15:17:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:12:07.008 15:17:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:12:07.008 15:17:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 84660' 00:12:07.008 killing process with pid 84660 00:12:07.008 15:17:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 84660 00:12:07.008 [2024-11-19 15:17:57.197211] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:07.008 15:17:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 84660 00:12:07.008 [2024-11-19 15:17:57.277015] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:07.268 15:17:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:12:07.268 00:12:07.268 real 0m9.751s 00:12:07.268 user 0m16.480s 00:12:07.268 sys 0m2.027s 00:12:07.268 15:17:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:12:07.268 15:17:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:07.268 ************************************ 00:12:07.268 END TEST raid_state_function_test_sb 00:12:07.268 ************************************ 00:12:07.529 15:17:57 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid1 4 00:12:07.529 15:17:57 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:12:07.529 15:17:57 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:12:07.529 15:17:57 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:07.529 ************************************ 00:12:07.529 START TEST raid_superblock_test 00:12:07.529 ************************************ 00:12:07.529 15:17:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid1 4 00:12:07.529 15:17:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:12:07.529 15:17:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:12:07.529 15:17:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:12:07.529 15:17:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:12:07.529 15:17:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:12:07.529 15:17:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:12:07.529 15:17:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:12:07.529 15:17:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:12:07.529 15:17:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:12:07.529 15:17:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:12:07.529 15:17:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:12:07.529 15:17:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:12:07.529 15:17:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:12:07.529 15:17:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:12:07.529 15:17:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:12:07.529 15:17:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=85314 00:12:07.529 15:17:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 85314 00:12:07.529 15:17:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:12:07.529 15:17:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 85314 ']' 00:12:07.529 15:17:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:07.529 15:17:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:12:07.529 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:07.529 15:17:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:07.529 15:17:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:12:07.529 15:17:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:07.529 [2024-11-19 15:17:57.749604] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:12:07.529 [2024-11-19 15:17:57.749833] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid85314 ] 00:12:07.789 [2024-11-19 15:17:57.906169] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:07.789 [2024-11-19 15:17:57.945797] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:12:07.789 [2024-11-19 15:17:58.020854] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:07.789 [2024-11-19 15:17:58.020897] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:08.357 15:17:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:12:08.357 15:17:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:12:08.357 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:12:08.357 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:12:08.357 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:12:08.357 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:12:08.357 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:12:08.357 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:12:08.357 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:12:08.357 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:12:08.357 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:12:08.357 15:17:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.357 15:17:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:08.357 malloc1 00:12:08.357 15:17:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.357 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:12:08.357 15:17:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.357 15:17:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:08.357 [2024-11-19 15:17:58.599859] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:12:08.357 [2024-11-19 15:17:58.600042] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:08.357 [2024-11-19 15:17:58.600095] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:12:08.357 [2024-11-19 15:17:58.600135] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:08.357 [2024-11-19 15:17:58.602721] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:08.357 [2024-11-19 15:17:58.602807] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:12:08.357 pt1 00:12:08.357 15:17:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.357 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:12:08.357 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:12:08.357 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:12:08.357 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:12:08.357 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:12:08.357 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:12:08.358 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:12:08.358 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:12:08.358 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:12:08.358 15:17:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.358 15:17:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:08.358 malloc2 00:12:08.358 15:17:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.358 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:12:08.358 15:17:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.358 15:17:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:08.358 [2024-11-19 15:17:58.639039] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:12:08.358 [2024-11-19 15:17:58.639122] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:08.358 [2024-11-19 15:17:58.639145] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:12:08.358 [2024-11-19 15:17:58.639157] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:08.358 [2024-11-19 15:17:58.641800] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:08.358 [2024-11-19 15:17:58.641925] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:12:08.358 pt2 00:12:08.358 15:17:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.358 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:12:08.358 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:12:08.358 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:12:08.358 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:12:08.358 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:12:08.358 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:12:08.358 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:12:08.358 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:12:08.358 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:12:08.358 15:17:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.358 15:17:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:08.358 malloc3 00:12:08.358 15:17:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.358 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:12:08.358 15:17:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.358 15:17:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:08.358 [2024-11-19 15:17:58.674244] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:12:08.358 [2024-11-19 15:17:58.674423] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:08.358 [2024-11-19 15:17:58.674470] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:12:08.358 [2024-11-19 15:17:58.674506] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:08.358 [2024-11-19 15:17:58.677182] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:08.358 [2024-11-19 15:17:58.677263] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:12:08.358 pt3 00:12:08.358 15:17:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.358 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:12:08.358 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:12:08.358 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:12:08.358 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:12:08.358 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:12:08.358 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:12:08.358 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:12:08.358 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:12:08.358 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:12:08.358 15:17:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.358 15:17:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:08.617 malloc4 00:12:08.617 15:17:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.617 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:12:08.617 15:17:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.617 15:17:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:08.617 [2024-11-19 15:17:58.717582] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:12:08.617 [2024-11-19 15:17:58.717744] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:08.617 [2024-11-19 15:17:58.717784] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:12:08.617 [2024-11-19 15:17:58.717819] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:08.617 [2024-11-19 15:17:58.720529] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:08.617 [2024-11-19 15:17:58.720618] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:12:08.617 pt4 00:12:08.617 15:17:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.617 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:12:08.617 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:12:08.617 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:12:08.617 15:17:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.617 15:17:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:08.617 [2024-11-19 15:17:58.729571] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:12:08.617 [2024-11-19 15:17:58.731867] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:12:08.617 [2024-11-19 15:17:58.732004] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:12:08.617 [2024-11-19 15:17:58.732082] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:12:08.617 [2024-11-19 15:17:58.732301] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:12:08.617 [2024-11-19 15:17:58.732352] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:08.617 [2024-11-19 15:17:58.732671] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:12:08.617 [2024-11-19 15:17:58.732908] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:12:08.617 [2024-11-19 15:17:58.732979] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:12:08.617 [2024-11-19 15:17:58.733164] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:08.617 15:17:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.617 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:12:08.617 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:08.617 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:08.617 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:08.617 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:08.617 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:08.617 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:08.617 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:08.617 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:08.617 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:08.617 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:08.617 15:17:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.617 15:17:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:08.617 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:08.617 15:17:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.617 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:08.617 "name": "raid_bdev1", 00:12:08.617 "uuid": "b030130f-934a-44e3-894e-27483fe25d3f", 00:12:08.617 "strip_size_kb": 0, 00:12:08.617 "state": "online", 00:12:08.617 "raid_level": "raid1", 00:12:08.617 "superblock": true, 00:12:08.617 "num_base_bdevs": 4, 00:12:08.617 "num_base_bdevs_discovered": 4, 00:12:08.617 "num_base_bdevs_operational": 4, 00:12:08.617 "base_bdevs_list": [ 00:12:08.617 { 00:12:08.617 "name": "pt1", 00:12:08.617 "uuid": "00000000-0000-0000-0000-000000000001", 00:12:08.617 "is_configured": true, 00:12:08.617 "data_offset": 2048, 00:12:08.617 "data_size": 63488 00:12:08.617 }, 00:12:08.617 { 00:12:08.617 "name": "pt2", 00:12:08.617 "uuid": "00000000-0000-0000-0000-000000000002", 00:12:08.617 "is_configured": true, 00:12:08.617 "data_offset": 2048, 00:12:08.617 "data_size": 63488 00:12:08.617 }, 00:12:08.617 { 00:12:08.617 "name": "pt3", 00:12:08.617 "uuid": "00000000-0000-0000-0000-000000000003", 00:12:08.617 "is_configured": true, 00:12:08.617 "data_offset": 2048, 00:12:08.617 "data_size": 63488 00:12:08.617 }, 00:12:08.617 { 00:12:08.617 "name": "pt4", 00:12:08.618 "uuid": "00000000-0000-0000-0000-000000000004", 00:12:08.618 "is_configured": true, 00:12:08.618 "data_offset": 2048, 00:12:08.618 "data_size": 63488 00:12:08.618 } 00:12:08.618 ] 00:12:08.618 }' 00:12:08.618 15:17:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:08.618 15:17:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:08.876 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:12:08.876 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:12:08.876 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:12:08.876 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:12:08.876 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:12:08.876 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:12:08.876 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:12:08.876 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:08.876 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.876 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:08.876 [2024-11-19 15:17:59.193185] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:09.136 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:09.136 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:12:09.136 "name": "raid_bdev1", 00:12:09.136 "aliases": [ 00:12:09.136 "b030130f-934a-44e3-894e-27483fe25d3f" 00:12:09.136 ], 00:12:09.136 "product_name": "Raid Volume", 00:12:09.136 "block_size": 512, 00:12:09.136 "num_blocks": 63488, 00:12:09.136 "uuid": "b030130f-934a-44e3-894e-27483fe25d3f", 00:12:09.136 "assigned_rate_limits": { 00:12:09.136 "rw_ios_per_sec": 0, 00:12:09.136 "rw_mbytes_per_sec": 0, 00:12:09.136 "r_mbytes_per_sec": 0, 00:12:09.136 "w_mbytes_per_sec": 0 00:12:09.136 }, 00:12:09.136 "claimed": false, 00:12:09.136 "zoned": false, 00:12:09.136 "supported_io_types": { 00:12:09.136 "read": true, 00:12:09.136 "write": true, 00:12:09.136 "unmap": false, 00:12:09.136 "flush": false, 00:12:09.136 "reset": true, 00:12:09.136 "nvme_admin": false, 00:12:09.136 "nvme_io": false, 00:12:09.136 "nvme_io_md": false, 00:12:09.136 "write_zeroes": true, 00:12:09.136 "zcopy": false, 00:12:09.136 "get_zone_info": false, 00:12:09.136 "zone_management": false, 00:12:09.136 "zone_append": false, 00:12:09.136 "compare": false, 00:12:09.136 "compare_and_write": false, 00:12:09.136 "abort": false, 00:12:09.136 "seek_hole": false, 00:12:09.136 "seek_data": false, 00:12:09.136 "copy": false, 00:12:09.136 "nvme_iov_md": false 00:12:09.136 }, 00:12:09.136 "memory_domains": [ 00:12:09.136 { 00:12:09.136 "dma_device_id": "system", 00:12:09.136 "dma_device_type": 1 00:12:09.136 }, 00:12:09.136 { 00:12:09.136 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:09.136 "dma_device_type": 2 00:12:09.136 }, 00:12:09.136 { 00:12:09.136 "dma_device_id": "system", 00:12:09.136 "dma_device_type": 1 00:12:09.136 }, 00:12:09.136 { 00:12:09.136 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:09.136 "dma_device_type": 2 00:12:09.136 }, 00:12:09.136 { 00:12:09.136 "dma_device_id": "system", 00:12:09.136 "dma_device_type": 1 00:12:09.136 }, 00:12:09.136 { 00:12:09.136 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:09.136 "dma_device_type": 2 00:12:09.136 }, 00:12:09.136 { 00:12:09.136 "dma_device_id": "system", 00:12:09.136 "dma_device_type": 1 00:12:09.136 }, 00:12:09.136 { 00:12:09.136 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:09.136 "dma_device_type": 2 00:12:09.136 } 00:12:09.136 ], 00:12:09.136 "driver_specific": { 00:12:09.136 "raid": { 00:12:09.136 "uuid": "b030130f-934a-44e3-894e-27483fe25d3f", 00:12:09.136 "strip_size_kb": 0, 00:12:09.136 "state": "online", 00:12:09.136 "raid_level": "raid1", 00:12:09.136 "superblock": true, 00:12:09.136 "num_base_bdevs": 4, 00:12:09.136 "num_base_bdevs_discovered": 4, 00:12:09.136 "num_base_bdevs_operational": 4, 00:12:09.136 "base_bdevs_list": [ 00:12:09.136 { 00:12:09.136 "name": "pt1", 00:12:09.136 "uuid": "00000000-0000-0000-0000-000000000001", 00:12:09.136 "is_configured": true, 00:12:09.136 "data_offset": 2048, 00:12:09.136 "data_size": 63488 00:12:09.136 }, 00:12:09.136 { 00:12:09.136 "name": "pt2", 00:12:09.136 "uuid": "00000000-0000-0000-0000-000000000002", 00:12:09.136 "is_configured": true, 00:12:09.136 "data_offset": 2048, 00:12:09.136 "data_size": 63488 00:12:09.136 }, 00:12:09.136 { 00:12:09.136 "name": "pt3", 00:12:09.136 "uuid": "00000000-0000-0000-0000-000000000003", 00:12:09.136 "is_configured": true, 00:12:09.136 "data_offset": 2048, 00:12:09.136 "data_size": 63488 00:12:09.136 }, 00:12:09.136 { 00:12:09.136 "name": "pt4", 00:12:09.136 "uuid": "00000000-0000-0000-0000-000000000004", 00:12:09.136 "is_configured": true, 00:12:09.136 "data_offset": 2048, 00:12:09.136 "data_size": 63488 00:12:09.136 } 00:12:09.136 ] 00:12:09.136 } 00:12:09.136 } 00:12:09.136 }' 00:12:09.136 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:12:09.136 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:12:09.136 pt2 00:12:09.136 pt3 00:12:09.136 pt4' 00:12:09.136 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:09.136 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:12:09.136 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:09.136 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:12:09.136 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:09.136 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:09.136 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:09.136 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:09.136 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:09.136 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:09.136 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:09.136 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:12:09.136 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:09.136 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:09.136 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:09.136 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:09.136 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:09.136 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:09.136 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:09.136 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:12:09.136 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:09.136 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:09.136 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:09.136 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:09.136 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:09.136 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:09.136 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:09.136 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:12:09.136 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:09.136 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:09.136 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:09.395 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:09.396 [2024-11-19 15:17:59.520528] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=b030130f-934a-44e3-894e-27483fe25d3f 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z b030130f-934a-44e3-894e-27483fe25d3f ']' 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:09.396 [2024-11-19 15:17:59.564138] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:09.396 [2024-11-19 15:17:59.564194] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:09.396 [2024-11-19 15:17:59.564325] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:09.396 [2024-11-19 15:17:59.564439] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:09.396 [2024-11-19 15:17:59.564450] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:09.396 [2024-11-19 15:17:59.719929] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:12:09.396 [2024-11-19 15:17:59.722347] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:12:09.396 [2024-11-19 15:17:59.722451] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:12:09.396 [2024-11-19 15:17:59.722519] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:12:09.396 [2024-11-19 15:17:59.722603] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:12:09.396 [2024-11-19 15:17:59.722693] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:12:09.396 [2024-11-19 15:17:59.722765] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:12:09.396 [2024-11-19 15:17:59.722820] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:12:09.396 [2024-11-19 15:17:59.722881] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:09.396 [2024-11-19 15:17:59.722932] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:12:09.396 request: 00:12:09.396 { 00:12:09.396 "name": "raid_bdev1", 00:12:09.396 "raid_level": "raid1", 00:12:09.396 "base_bdevs": [ 00:12:09.396 "malloc1", 00:12:09.396 "malloc2", 00:12:09.396 "malloc3", 00:12:09.396 "malloc4" 00:12:09.396 ], 00:12:09.396 "superblock": false, 00:12:09.396 "method": "bdev_raid_create", 00:12:09.396 "req_id": 1 00:12:09.396 } 00:12:09.396 Got JSON-RPC error response 00:12:09.396 response: 00:12:09.396 { 00:12:09.396 "code": -17, 00:12:09.396 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:12:09.396 } 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:09.396 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:12:09.656 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:09.656 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:12:09.656 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:12:09.656 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:12:09.656 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:09.656 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:09.656 [2024-11-19 15:17:59.779841] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:12:09.656 [2024-11-19 15:17:59.779995] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:09.656 [2024-11-19 15:17:59.780028] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:12:09.656 [2024-11-19 15:17:59.780039] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:09.656 [2024-11-19 15:17:59.782661] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:09.656 [2024-11-19 15:17:59.782699] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:12:09.656 [2024-11-19 15:17:59.782792] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:12:09.656 [2024-11-19 15:17:59.782845] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:12:09.656 pt1 00:12:09.656 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:09.656 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 4 00:12:09.656 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:09.656 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:09.656 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:09.656 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:09.656 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:09.656 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:09.656 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:09.656 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:09.656 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:09.656 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:09.656 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:09.656 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:09.656 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:09.656 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:09.656 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:09.656 "name": "raid_bdev1", 00:12:09.656 "uuid": "b030130f-934a-44e3-894e-27483fe25d3f", 00:12:09.656 "strip_size_kb": 0, 00:12:09.656 "state": "configuring", 00:12:09.656 "raid_level": "raid1", 00:12:09.656 "superblock": true, 00:12:09.656 "num_base_bdevs": 4, 00:12:09.656 "num_base_bdevs_discovered": 1, 00:12:09.656 "num_base_bdevs_operational": 4, 00:12:09.656 "base_bdevs_list": [ 00:12:09.656 { 00:12:09.656 "name": "pt1", 00:12:09.656 "uuid": "00000000-0000-0000-0000-000000000001", 00:12:09.656 "is_configured": true, 00:12:09.656 "data_offset": 2048, 00:12:09.656 "data_size": 63488 00:12:09.656 }, 00:12:09.656 { 00:12:09.656 "name": null, 00:12:09.656 "uuid": "00000000-0000-0000-0000-000000000002", 00:12:09.656 "is_configured": false, 00:12:09.656 "data_offset": 2048, 00:12:09.656 "data_size": 63488 00:12:09.656 }, 00:12:09.656 { 00:12:09.656 "name": null, 00:12:09.656 "uuid": "00000000-0000-0000-0000-000000000003", 00:12:09.656 "is_configured": false, 00:12:09.656 "data_offset": 2048, 00:12:09.656 "data_size": 63488 00:12:09.656 }, 00:12:09.656 { 00:12:09.656 "name": null, 00:12:09.656 "uuid": "00000000-0000-0000-0000-000000000004", 00:12:09.656 "is_configured": false, 00:12:09.656 "data_offset": 2048, 00:12:09.656 "data_size": 63488 00:12:09.656 } 00:12:09.656 ] 00:12:09.656 }' 00:12:09.656 15:17:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:09.656 15:17:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:09.916 15:18:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:12:09.916 15:18:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:12:09.916 15:18:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:09.916 15:18:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:09.916 [2024-11-19 15:18:00.251095] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:12:09.916 [2024-11-19 15:18:00.251271] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:09.916 [2024-11-19 15:18:00.251317] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:12:09.916 [2024-11-19 15:18:00.251330] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:09.916 [2024-11-19 15:18:00.251892] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:09.916 [2024-11-19 15:18:00.251915] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:12:09.916 [2024-11-19 15:18:00.252035] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:12:09.916 [2024-11-19 15:18:00.252062] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:12:10.175 pt2 00:12:10.175 15:18:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:10.175 15:18:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:12:10.175 15:18:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:10.175 15:18:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:10.175 [2024-11-19 15:18:00.263062] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:12:10.175 15:18:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:10.175 15:18:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 4 00:12:10.175 15:18:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:10.175 15:18:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:10.175 15:18:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:10.175 15:18:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:10.175 15:18:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:10.175 15:18:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:10.175 15:18:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:10.175 15:18:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:10.175 15:18:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:10.175 15:18:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:10.175 15:18:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:10.175 15:18:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:10.175 15:18:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:10.175 15:18:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:10.175 15:18:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:10.175 "name": "raid_bdev1", 00:12:10.175 "uuid": "b030130f-934a-44e3-894e-27483fe25d3f", 00:12:10.175 "strip_size_kb": 0, 00:12:10.175 "state": "configuring", 00:12:10.175 "raid_level": "raid1", 00:12:10.175 "superblock": true, 00:12:10.175 "num_base_bdevs": 4, 00:12:10.175 "num_base_bdevs_discovered": 1, 00:12:10.175 "num_base_bdevs_operational": 4, 00:12:10.175 "base_bdevs_list": [ 00:12:10.175 { 00:12:10.175 "name": "pt1", 00:12:10.175 "uuid": "00000000-0000-0000-0000-000000000001", 00:12:10.175 "is_configured": true, 00:12:10.175 "data_offset": 2048, 00:12:10.175 "data_size": 63488 00:12:10.175 }, 00:12:10.175 { 00:12:10.175 "name": null, 00:12:10.175 "uuid": "00000000-0000-0000-0000-000000000002", 00:12:10.175 "is_configured": false, 00:12:10.175 "data_offset": 0, 00:12:10.175 "data_size": 63488 00:12:10.175 }, 00:12:10.175 { 00:12:10.175 "name": null, 00:12:10.175 "uuid": "00000000-0000-0000-0000-000000000003", 00:12:10.175 "is_configured": false, 00:12:10.175 "data_offset": 2048, 00:12:10.175 "data_size": 63488 00:12:10.175 }, 00:12:10.175 { 00:12:10.175 "name": null, 00:12:10.175 "uuid": "00000000-0000-0000-0000-000000000004", 00:12:10.175 "is_configured": false, 00:12:10.175 "data_offset": 2048, 00:12:10.175 "data_size": 63488 00:12:10.175 } 00:12:10.175 ] 00:12:10.175 }' 00:12:10.175 15:18:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:10.175 15:18:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:10.435 15:18:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:12:10.435 15:18:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:12:10.435 15:18:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:12:10.435 15:18:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:10.435 15:18:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:10.435 [2024-11-19 15:18:00.714245] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:12:10.435 [2024-11-19 15:18:00.714395] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:10.435 [2024-11-19 15:18:00.714434] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:12:10.435 [2024-11-19 15:18:00.714465] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:10.435 [2024-11-19 15:18:00.714982] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:10.435 [2024-11-19 15:18:00.715061] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:12:10.435 [2024-11-19 15:18:00.715179] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:12:10.435 [2024-11-19 15:18:00.715235] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:12:10.435 pt2 00:12:10.435 15:18:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:10.435 15:18:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:12:10.435 15:18:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:12:10.435 15:18:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:12:10.435 15:18:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:10.435 15:18:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:10.435 [2024-11-19 15:18:00.722184] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:12:10.435 [2024-11-19 15:18:00.722271] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:10.435 [2024-11-19 15:18:00.722303] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:12:10.435 [2024-11-19 15:18:00.722366] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:10.435 [2024-11-19 15:18:00.722764] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:10.435 [2024-11-19 15:18:00.722826] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:12:10.435 [2024-11-19 15:18:00.722907] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:12:10.435 [2024-11-19 15:18:00.722965] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:12:10.435 pt3 00:12:10.435 15:18:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:10.435 15:18:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:12:10.435 15:18:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:12:10.435 15:18:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:12:10.435 15:18:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:10.435 15:18:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:10.435 [2024-11-19 15:18:00.730177] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:12:10.435 [2024-11-19 15:18:00.730270] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:10.435 [2024-11-19 15:18:00.730300] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:12:10.435 [2024-11-19 15:18:00.730328] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:10.435 [2024-11-19 15:18:00.730645] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:10.435 [2024-11-19 15:18:00.730703] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:12:10.435 [2024-11-19 15:18:00.730778] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:12:10.435 [2024-11-19 15:18:00.730822] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:12:10.435 [2024-11-19 15:18:00.730952] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:12:10.435 [2024-11-19 15:18:00.731018] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:10.435 [2024-11-19 15:18:00.731282] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:12:10.435 [2024-11-19 15:18:00.731464] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:12:10.435 [2024-11-19 15:18:00.731513] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:12:10.435 [2024-11-19 15:18:00.731695] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:10.435 pt4 00:12:10.435 15:18:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:10.435 15:18:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:12:10.435 15:18:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:12:10.435 15:18:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:12:10.435 15:18:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:10.435 15:18:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:10.435 15:18:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:10.435 15:18:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:10.435 15:18:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:10.435 15:18:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:10.435 15:18:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:10.435 15:18:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:10.435 15:18:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:10.435 15:18:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:10.435 15:18:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:10.435 15:18:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:10.435 15:18:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:10.435 15:18:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:10.435 15:18:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:10.435 "name": "raid_bdev1", 00:12:10.435 "uuid": "b030130f-934a-44e3-894e-27483fe25d3f", 00:12:10.435 "strip_size_kb": 0, 00:12:10.435 "state": "online", 00:12:10.435 "raid_level": "raid1", 00:12:10.435 "superblock": true, 00:12:10.435 "num_base_bdevs": 4, 00:12:10.435 "num_base_bdevs_discovered": 4, 00:12:10.435 "num_base_bdevs_operational": 4, 00:12:10.435 "base_bdevs_list": [ 00:12:10.435 { 00:12:10.435 "name": "pt1", 00:12:10.435 "uuid": "00000000-0000-0000-0000-000000000001", 00:12:10.435 "is_configured": true, 00:12:10.435 "data_offset": 2048, 00:12:10.435 "data_size": 63488 00:12:10.435 }, 00:12:10.435 { 00:12:10.435 "name": "pt2", 00:12:10.435 "uuid": "00000000-0000-0000-0000-000000000002", 00:12:10.435 "is_configured": true, 00:12:10.435 "data_offset": 2048, 00:12:10.435 "data_size": 63488 00:12:10.435 }, 00:12:10.435 { 00:12:10.435 "name": "pt3", 00:12:10.435 "uuid": "00000000-0000-0000-0000-000000000003", 00:12:10.435 "is_configured": true, 00:12:10.435 "data_offset": 2048, 00:12:10.435 "data_size": 63488 00:12:10.435 }, 00:12:10.435 { 00:12:10.435 "name": "pt4", 00:12:10.435 "uuid": "00000000-0000-0000-0000-000000000004", 00:12:10.435 "is_configured": true, 00:12:10.435 "data_offset": 2048, 00:12:10.435 "data_size": 63488 00:12:10.435 } 00:12:10.435 ] 00:12:10.435 }' 00:12:10.435 15:18:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:10.435 15:18:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:11.005 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:12:11.005 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:12:11.005 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:12:11.005 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:12:11.005 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:12:11.005 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:12:11.005 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:11.005 15:18:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.005 15:18:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:11.005 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:12:11.005 [2024-11-19 15:18:01.149749] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:11.005 15:18:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.005 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:12:11.005 "name": "raid_bdev1", 00:12:11.005 "aliases": [ 00:12:11.005 "b030130f-934a-44e3-894e-27483fe25d3f" 00:12:11.005 ], 00:12:11.005 "product_name": "Raid Volume", 00:12:11.005 "block_size": 512, 00:12:11.005 "num_blocks": 63488, 00:12:11.005 "uuid": "b030130f-934a-44e3-894e-27483fe25d3f", 00:12:11.005 "assigned_rate_limits": { 00:12:11.005 "rw_ios_per_sec": 0, 00:12:11.005 "rw_mbytes_per_sec": 0, 00:12:11.005 "r_mbytes_per_sec": 0, 00:12:11.005 "w_mbytes_per_sec": 0 00:12:11.005 }, 00:12:11.005 "claimed": false, 00:12:11.005 "zoned": false, 00:12:11.005 "supported_io_types": { 00:12:11.005 "read": true, 00:12:11.005 "write": true, 00:12:11.005 "unmap": false, 00:12:11.005 "flush": false, 00:12:11.005 "reset": true, 00:12:11.005 "nvme_admin": false, 00:12:11.005 "nvme_io": false, 00:12:11.005 "nvme_io_md": false, 00:12:11.005 "write_zeroes": true, 00:12:11.005 "zcopy": false, 00:12:11.005 "get_zone_info": false, 00:12:11.005 "zone_management": false, 00:12:11.005 "zone_append": false, 00:12:11.005 "compare": false, 00:12:11.005 "compare_and_write": false, 00:12:11.005 "abort": false, 00:12:11.005 "seek_hole": false, 00:12:11.005 "seek_data": false, 00:12:11.005 "copy": false, 00:12:11.005 "nvme_iov_md": false 00:12:11.005 }, 00:12:11.005 "memory_domains": [ 00:12:11.005 { 00:12:11.005 "dma_device_id": "system", 00:12:11.005 "dma_device_type": 1 00:12:11.005 }, 00:12:11.005 { 00:12:11.005 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:11.005 "dma_device_type": 2 00:12:11.005 }, 00:12:11.005 { 00:12:11.005 "dma_device_id": "system", 00:12:11.005 "dma_device_type": 1 00:12:11.005 }, 00:12:11.005 { 00:12:11.005 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:11.005 "dma_device_type": 2 00:12:11.005 }, 00:12:11.005 { 00:12:11.005 "dma_device_id": "system", 00:12:11.005 "dma_device_type": 1 00:12:11.005 }, 00:12:11.005 { 00:12:11.005 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:11.005 "dma_device_type": 2 00:12:11.005 }, 00:12:11.005 { 00:12:11.005 "dma_device_id": "system", 00:12:11.005 "dma_device_type": 1 00:12:11.005 }, 00:12:11.005 { 00:12:11.005 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:12:11.005 "dma_device_type": 2 00:12:11.005 } 00:12:11.005 ], 00:12:11.005 "driver_specific": { 00:12:11.005 "raid": { 00:12:11.005 "uuid": "b030130f-934a-44e3-894e-27483fe25d3f", 00:12:11.005 "strip_size_kb": 0, 00:12:11.005 "state": "online", 00:12:11.005 "raid_level": "raid1", 00:12:11.005 "superblock": true, 00:12:11.005 "num_base_bdevs": 4, 00:12:11.005 "num_base_bdevs_discovered": 4, 00:12:11.005 "num_base_bdevs_operational": 4, 00:12:11.005 "base_bdevs_list": [ 00:12:11.005 { 00:12:11.005 "name": "pt1", 00:12:11.005 "uuid": "00000000-0000-0000-0000-000000000001", 00:12:11.005 "is_configured": true, 00:12:11.005 "data_offset": 2048, 00:12:11.005 "data_size": 63488 00:12:11.005 }, 00:12:11.005 { 00:12:11.005 "name": "pt2", 00:12:11.005 "uuid": "00000000-0000-0000-0000-000000000002", 00:12:11.005 "is_configured": true, 00:12:11.005 "data_offset": 2048, 00:12:11.005 "data_size": 63488 00:12:11.005 }, 00:12:11.005 { 00:12:11.005 "name": "pt3", 00:12:11.005 "uuid": "00000000-0000-0000-0000-000000000003", 00:12:11.005 "is_configured": true, 00:12:11.005 "data_offset": 2048, 00:12:11.005 "data_size": 63488 00:12:11.005 }, 00:12:11.005 { 00:12:11.005 "name": "pt4", 00:12:11.005 "uuid": "00000000-0000-0000-0000-000000000004", 00:12:11.005 "is_configured": true, 00:12:11.005 "data_offset": 2048, 00:12:11.005 "data_size": 63488 00:12:11.005 } 00:12:11.005 ] 00:12:11.005 } 00:12:11.005 } 00:12:11.005 }' 00:12:11.005 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:12:11.005 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:12:11.005 pt2 00:12:11.005 pt3 00:12:11.005 pt4' 00:12:11.005 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:11.005 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:12:11.005 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:11.005 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:12:11.005 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:11.005 15:18:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.005 15:18:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:11.005 15:18:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.006 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:11.006 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:11.006 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:11.006 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:12:11.006 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:11.006 15:18:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.006 15:18:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:11.006 15:18:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.266 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:11.266 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:11.266 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:11.266 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:11.266 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:12:11.266 15:18:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.266 15:18:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:11.266 15:18:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.266 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:11.266 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:11.266 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:12:11.266 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:12:11.266 15:18:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.266 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:12:11.266 15:18:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:11.266 15:18:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.266 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:12:11.266 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:12:11.266 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:11.266 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:12:11.266 15:18:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.266 15:18:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:11.266 [2024-11-19 15:18:01.453308] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:11.266 15:18:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.266 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' b030130f-934a-44e3-894e-27483fe25d3f '!=' b030130f-934a-44e3-894e-27483fe25d3f ']' 00:12:11.266 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:12:11.266 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:12:11.266 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:12:11.266 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:12:11.266 15:18:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.266 15:18:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:11.266 [2024-11-19 15:18:01.497077] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:12:11.266 15:18:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.266 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:11.266 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:11.266 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:11.266 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:11.266 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:11.266 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:11.266 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:11.266 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:11.266 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:11.266 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:11.266 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:11.266 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:11.266 15:18:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.267 15:18:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:11.267 15:18:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.267 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:11.267 "name": "raid_bdev1", 00:12:11.267 "uuid": "b030130f-934a-44e3-894e-27483fe25d3f", 00:12:11.267 "strip_size_kb": 0, 00:12:11.267 "state": "online", 00:12:11.267 "raid_level": "raid1", 00:12:11.267 "superblock": true, 00:12:11.267 "num_base_bdevs": 4, 00:12:11.267 "num_base_bdevs_discovered": 3, 00:12:11.267 "num_base_bdevs_operational": 3, 00:12:11.267 "base_bdevs_list": [ 00:12:11.267 { 00:12:11.267 "name": null, 00:12:11.267 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:11.267 "is_configured": false, 00:12:11.267 "data_offset": 0, 00:12:11.267 "data_size": 63488 00:12:11.267 }, 00:12:11.267 { 00:12:11.267 "name": "pt2", 00:12:11.267 "uuid": "00000000-0000-0000-0000-000000000002", 00:12:11.267 "is_configured": true, 00:12:11.267 "data_offset": 2048, 00:12:11.267 "data_size": 63488 00:12:11.267 }, 00:12:11.267 { 00:12:11.267 "name": "pt3", 00:12:11.267 "uuid": "00000000-0000-0000-0000-000000000003", 00:12:11.267 "is_configured": true, 00:12:11.267 "data_offset": 2048, 00:12:11.267 "data_size": 63488 00:12:11.267 }, 00:12:11.267 { 00:12:11.267 "name": "pt4", 00:12:11.267 "uuid": "00000000-0000-0000-0000-000000000004", 00:12:11.267 "is_configured": true, 00:12:11.267 "data_offset": 2048, 00:12:11.267 "data_size": 63488 00:12:11.267 } 00:12:11.267 ] 00:12:11.267 }' 00:12:11.267 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:11.267 15:18:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:11.838 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:12:11.838 15:18:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.838 15:18:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:11.838 [2024-11-19 15:18:01.908244] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:11.838 [2024-11-19 15:18:01.908292] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:11.838 [2024-11-19 15:18:01.908412] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:11.838 [2024-11-19 15:18:01.908493] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:11.838 [2024-11-19 15:18:01.908507] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:12:11.838 15:18:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.838 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:11.838 15:18:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.838 15:18:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:11.838 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:12:11.838 15:18:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.838 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:12:11.838 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:12:11.838 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:12:11.838 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:12:11.838 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:12:11.838 15:18:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.838 15:18:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:11.838 15:18:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.838 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:12:11.838 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:12:11.838 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:12:11.838 15:18:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.838 15:18:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:11.838 15:18:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.838 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:12:11.838 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:12:11.838 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt4 00:12:11.838 15:18:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.838 15:18:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:11.838 15:18:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.838 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:12:11.838 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:12:11.838 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:12:11.838 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:12:11.838 15:18:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:12:11.838 15:18:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.838 15:18:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:11.838 [2024-11-19 15:18:02.004111] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:12:11.838 [2024-11-19 15:18:02.004207] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:11.838 [2024-11-19 15:18:02.004228] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:12:11.838 [2024-11-19 15:18:02.004241] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:11.838 [2024-11-19 15:18:02.006751] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:11.838 [2024-11-19 15:18:02.006795] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:12:11.838 [2024-11-19 15:18:02.006900] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:12:11.838 [2024-11-19 15:18:02.006944] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:12:11.838 pt2 00:12:11.838 15:18:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.838 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:12:11.838 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:11.838 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:11.838 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:11.838 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:11.838 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:11.838 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:11.838 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:11.838 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:11.838 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:11.838 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:11.838 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:11.838 15:18:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.838 15:18:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:11.838 15:18:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.838 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:11.838 "name": "raid_bdev1", 00:12:11.838 "uuid": "b030130f-934a-44e3-894e-27483fe25d3f", 00:12:11.838 "strip_size_kb": 0, 00:12:11.838 "state": "configuring", 00:12:11.838 "raid_level": "raid1", 00:12:11.838 "superblock": true, 00:12:11.838 "num_base_bdevs": 4, 00:12:11.838 "num_base_bdevs_discovered": 1, 00:12:11.838 "num_base_bdevs_operational": 3, 00:12:11.838 "base_bdevs_list": [ 00:12:11.838 { 00:12:11.838 "name": null, 00:12:11.838 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:11.838 "is_configured": false, 00:12:11.838 "data_offset": 2048, 00:12:11.838 "data_size": 63488 00:12:11.838 }, 00:12:11.838 { 00:12:11.838 "name": "pt2", 00:12:11.838 "uuid": "00000000-0000-0000-0000-000000000002", 00:12:11.838 "is_configured": true, 00:12:11.838 "data_offset": 2048, 00:12:11.838 "data_size": 63488 00:12:11.838 }, 00:12:11.838 { 00:12:11.839 "name": null, 00:12:11.839 "uuid": "00000000-0000-0000-0000-000000000003", 00:12:11.839 "is_configured": false, 00:12:11.839 "data_offset": 2048, 00:12:11.839 "data_size": 63488 00:12:11.839 }, 00:12:11.839 { 00:12:11.839 "name": null, 00:12:11.839 "uuid": "00000000-0000-0000-0000-000000000004", 00:12:11.839 "is_configured": false, 00:12:11.839 "data_offset": 2048, 00:12:11.839 "data_size": 63488 00:12:11.839 } 00:12:11.839 ] 00:12:11.839 }' 00:12:11.839 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:11.839 15:18:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:12.099 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:12:12.099 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:12:12.099 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:12:12.099 15:18:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:12.099 15:18:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:12.099 [2024-11-19 15:18:02.407645] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:12:12.099 [2024-11-19 15:18:02.407764] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:12.099 [2024-11-19 15:18:02.407788] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:12:12.099 [2024-11-19 15:18:02.407803] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:12.099 [2024-11-19 15:18:02.408281] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:12.099 [2024-11-19 15:18:02.408310] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:12:12.099 [2024-11-19 15:18:02.408403] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:12:12.099 [2024-11-19 15:18:02.408446] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:12:12.099 pt3 00:12:12.099 15:18:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:12.099 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:12:12.099 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:12.099 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:12.099 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:12.099 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:12.099 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:12.099 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:12.099 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:12.099 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:12.099 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:12.099 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:12.099 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:12.099 15:18:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:12.099 15:18:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:12.359 15:18:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:12.359 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:12.359 "name": "raid_bdev1", 00:12:12.359 "uuid": "b030130f-934a-44e3-894e-27483fe25d3f", 00:12:12.359 "strip_size_kb": 0, 00:12:12.359 "state": "configuring", 00:12:12.359 "raid_level": "raid1", 00:12:12.359 "superblock": true, 00:12:12.359 "num_base_bdevs": 4, 00:12:12.359 "num_base_bdevs_discovered": 2, 00:12:12.359 "num_base_bdevs_operational": 3, 00:12:12.359 "base_bdevs_list": [ 00:12:12.359 { 00:12:12.359 "name": null, 00:12:12.359 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:12.359 "is_configured": false, 00:12:12.359 "data_offset": 2048, 00:12:12.359 "data_size": 63488 00:12:12.359 }, 00:12:12.359 { 00:12:12.359 "name": "pt2", 00:12:12.359 "uuid": "00000000-0000-0000-0000-000000000002", 00:12:12.359 "is_configured": true, 00:12:12.359 "data_offset": 2048, 00:12:12.359 "data_size": 63488 00:12:12.359 }, 00:12:12.359 { 00:12:12.359 "name": "pt3", 00:12:12.359 "uuid": "00000000-0000-0000-0000-000000000003", 00:12:12.359 "is_configured": true, 00:12:12.359 "data_offset": 2048, 00:12:12.359 "data_size": 63488 00:12:12.359 }, 00:12:12.359 { 00:12:12.359 "name": null, 00:12:12.359 "uuid": "00000000-0000-0000-0000-000000000004", 00:12:12.359 "is_configured": false, 00:12:12.359 "data_offset": 2048, 00:12:12.359 "data_size": 63488 00:12:12.359 } 00:12:12.359 ] 00:12:12.359 }' 00:12:12.359 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:12.359 15:18:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:12.619 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:12:12.619 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:12:12.619 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@519 -- # i=3 00:12:12.619 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:12:12.619 15:18:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:12.619 15:18:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:12.619 [2024-11-19 15:18:02.850829] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:12:12.619 [2024-11-19 15:18:02.850927] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:12.619 [2024-11-19 15:18:02.850951] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:12:12.619 [2024-11-19 15:18:02.850978] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:12.619 [2024-11-19 15:18:02.851521] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:12.619 [2024-11-19 15:18:02.851558] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:12:12.619 [2024-11-19 15:18:02.851694] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:12:12.619 [2024-11-19 15:18:02.851730] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:12:12.619 [2024-11-19 15:18:02.851852] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:12:12.619 [2024-11-19 15:18:02.851869] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:12.619 [2024-11-19 15:18:02.852175] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:12:12.619 [2024-11-19 15:18:02.852340] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:12:12.619 [2024-11-19 15:18:02.852356] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:12:12.619 [2024-11-19 15:18:02.852475] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:12.619 pt4 00:12:12.619 15:18:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:12.619 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:12.619 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:12.620 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:12.620 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:12.620 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:12.620 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:12.620 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:12.620 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:12.620 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:12.620 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:12.620 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:12.620 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:12.620 15:18:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:12.620 15:18:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:12.620 15:18:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:12.620 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:12.620 "name": "raid_bdev1", 00:12:12.620 "uuid": "b030130f-934a-44e3-894e-27483fe25d3f", 00:12:12.620 "strip_size_kb": 0, 00:12:12.620 "state": "online", 00:12:12.620 "raid_level": "raid1", 00:12:12.620 "superblock": true, 00:12:12.620 "num_base_bdevs": 4, 00:12:12.620 "num_base_bdevs_discovered": 3, 00:12:12.620 "num_base_bdevs_operational": 3, 00:12:12.620 "base_bdevs_list": [ 00:12:12.620 { 00:12:12.620 "name": null, 00:12:12.620 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:12.620 "is_configured": false, 00:12:12.620 "data_offset": 2048, 00:12:12.620 "data_size": 63488 00:12:12.620 }, 00:12:12.620 { 00:12:12.620 "name": "pt2", 00:12:12.620 "uuid": "00000000-0000-0000-0000-000000000002", 00:12:12.620 "is_configured": true, 00:12:12.620 "data_offset": 2048, 00:12:12.620 "data_size": 63488 00:12:12.620 }, 00:12:12.620 { 00:12:12.620 "name": "pt3", 00:12:12.620 "uuid": "00000000-0000-0000-0000-000000000003", 00:12:12.620 "is_configured": true, 00:12:12.620 "data_offset": 2048, 00:12:12.620 "data_size": 63488 00:12:12.620 }, 00:12:12.620 { 00:12:12.620 "name": "pt4", 00:12:12.620 "uuid": "00000000-0000-0000-0000-000000000004", 00:12:12.620 "is_configured": true, 00:12:12.620 "data_offset": 2048, 00:12:12.620 "data_size": 63488 00:12:12.620 } 00:12:12.620 ] 00:12:12.620 }' 00:12:12.620 15:18:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:12.620 15:18:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:13.190 15:18:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:12:13.190 15:18:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:13.190 15:18:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:13.190 [2024-11-19 15:18:03.238153] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:13.190 [2024-11-19 15:18:03.238205] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:13.190 [2024-11-19 15:18:03.238303] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:13.190 [2024-11-19 15:18:03.238390] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:13.190 [2024-11-19 15:18:03.238401] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:12:13.190 15:18:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:13.190 15:18:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:13.190 15:18:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:13.190 15:18:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:12:13.190 15:18:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:13.190 15:18:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:13.190 15:18:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:12:13.190 15:18:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:12:13.190 15:18:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 4 -gt 2 ']' 00:12:13.190 15:18:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@534 -- # i=3 00:12:13.191 15:18:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt4 00:12:13.191 15:18:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:13.191 15:18:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:13.191 15:18:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:13.191 15:18:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:12:13.191 15:18:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:13.191 15:18:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:13.191 [2024-11-19 15:18:03.302037] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:12:13.191 [2024-11-19 15:18:03.302112] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:13.191 [2024-11-19 15:18:03.302138] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b480 00:12:13.191 [2024-11-19 15:18:03.302148] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:13.191 [2024-11-19 15:18:03.304568] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:13.191 [2024-11-19 15:18:03.304603] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:12:13.191 [2024-11-19 15:18:03.304683] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:12:13.191 [2024-11-19 15:18:03.304728] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:12:13.191 [2024-11-19 15:18:03.304859] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:12:13.191 [2024-11-19 15:18:03.304879] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:13.191 [2024-11-19 15:18:03.304898] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:12:13.191 [2024-11-19 15:18:03.304940] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:12:13.191 [2024-11-19 15:18:03.305064] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:12:13.191 pt1 00:12:13.191 15:18:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:13.191 15:18:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 4 -gt 2 ']' 00:12:13.191 15:18:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:12:13.191 15:18:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:13.191 15:18:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:12:13.191 15:18:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:13.191 15:18:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:13.191 15:18:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:13.191 15:18:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:13.191 15:18:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:13.191 15:18:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:13.191 15:18:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:13.191 15:18:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:13.191 15:18:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:13.191 15:18:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:13.191 15:18:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:13.191 15:18:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:13.191 15:18:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:13.191 "name": "raid_bdev1", 00:12:13.191 "uuid": "b030130f-934a-44e3-894e-27483fe25d3f", 00:12:13.191 "strip_size_kb": 0, 00:12:13.191 "state": "configuring", 00:12:13.191 "raid_level": "raid1", 00:12:13.191 "superblock": true, 00:12:13.191 "num_base_bdevs": 4, 00:12:13.191 "num_base_bdevs_discovered": 2, 00:12:13.191 "num_base_bdevs_operational": 3, 00:12:13.191 "base_bdevs_list": [ 00:12:13.191 { 00:12:13.191 "name": null, 00:12:13.191 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:13.191 "is_configured": false, 00:12:13.191 "data_offset": 2048, 00:12:13.191 "data_size": 63488 00:12:13.191 }, 00:12:13.191 { 00:12:13.191 "name": "pt2", 00:12:13.191 "uuid": "00000000-0000-0000-0000-000000000002", 00:12:13.191 "is_configured": true, 00:12:13.191 "data_offset": 2048, 00:12:13.191 "data_size": 63488 00:12:13.191 }, 00:12:13.191 { 00:12:13.191 "name": "pt3", 00:12:13.191 "uuid": "00000000-0000-0000-0000-000000000003", 00:12:13.191 "is_configured": true, 00:12:13.191 "data_offset": 2048, 00:12:13.191 "data_size": 63488 00:12:13.191 }, 00:12:13.191 { 00:12:13.191 "name": null, 00:12:13.191 "uuid": "00000000-0000-0000-0000-000000000004", 00:12:13.191 "is_configured": false, 00:12:13.191 "data_offset": 2048, 00:12:13.191 "data_size": 63488 00:12:13.191 } 00:12:13.191 ] 00:12:13.191 }' 00:12:13.191 15:18:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:13.191 15:18:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:13.450 15:18:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:12:13.450 15:18:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:13.450 15:18:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:13.450 15:18:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:12:13.450 15:18:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:13.450 15:18:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:12:13.450 15:18:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:12:13.450 15:18:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:13.450 15:18:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:13.710 [2024-11-19 15:18:03.789186] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:12:13.710 [2024-11-19 15:18:03.789274] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:13.710 [2024-11-19 15:18:03.789299] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ba80 00:12:13.710 [2024-11-19 15:18:03.789312] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:13.710 [2024-11-19 15:18:03.789808] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:13.710 [2024-11-19 15:18:03.789836] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:12:13.710 [2024-11-19 15:18:03.789922] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:12:13.710 [2024-11-19 15:18:03.789954] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:12:13.710 [2024-11-19 15:18:03.790080] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:12:13.710 [2024-11-19 15:18:03.790101] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:13.710 [2024-11-19 15:18:03.790389] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:12:13.710 [2024-11-19 15:18:03.790524] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:12:13.710 [2024-11-19 15:18:03.790537] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:12:13.710 [2024-11-19 15:18:03.790655] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:13.710 pt4 00:12:13.710 15:18:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:13.710 15:18:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:13.710 15:18:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:13.710 15:18:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:13.710 15:18:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:13.710 15:18:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:13.710 15:18:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:13.710 15:18:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:13.710 15:18:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:13.710 15:18:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:13.710 15:18:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:13.710 15:18:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:13.710 15:18:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:13.710 15:18:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:13.710 15:18:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:13.710 15:18:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:13.710 15:18:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:13.710 "name": "raid_bdev1", 00:12:13.710 "uuid": "b030130f-934a-44e3-894e-27483fe25d3f", 00:12:13.710 "strip_size_kb": 0, 00:12:13.710 "state": "online", 00:12:13.710 "raid_level": "raid1", 00:12:13.710 "superblock": true, 00:12:13.710 "num_base_bdevs": 4, 00:12:13.710 "num_base_bdevs_discovered": 3, 00:12:13.710 "num_base_bdevs_operational": 3, 00:12:13.710 "base_bdevs_list": [ 00:12:13.710 { 00:12:13.710 "name": null, 00:12:13.710 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:13.710 "is_configured": false, 00:12:13.710 "data_offset": 2048, 00:12:13.710 "data_size": 63488 00:12:13.710 }, 00:12:13.710 { 00:12:13.710 "name": "pt2", 00:12:13.710 "uuid": "00000000-0000-0000-0000-000000000002", 00:12:13.710 "is_configured": true, 00:12:13.710 "data_offset": 2048, 00:12:13.710 "data_size": 63488 00:12:13.710 }, 00:12:13.710 { 00:12:13.710 "name": "pt3", 00:12:13.710 "uuid": "00000000-0000-0000-0000-000000000003", 00:12:13.710 "is_configured": true, 00:12:13.710 "data_offset": 2048, 00:12:13.710 "data_size": 63488 00:12:13.710 }, 00:12:13.710 { 00:12:13.710 "name": "pt4", 00:12:13.710 "uuid": "00000000-0000-0000-0000-000000000004", 00:12:13.710 "is_configured": true, 00:12:13.710 "data_offset": 2048, 00:12:13.710 "data_size": 63488 00:12:13.710 } 00:12:13.710 ] 00:12:13.710 }' 00:12:13.710 15:18:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:13.710 15:18:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:13.970 15:18:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:12:13.970 15:18:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:12:13.970 15:18:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:13.970 15:18:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:13.970 15:18:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:13.970 15:18:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:12:13.970 15:18:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:13.970 15:18:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:12:13.970 15:18:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:13.970 15:18:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:13.970 [2024-11-19 15:18:04.236720] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:13.970 15:18:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:13.970 15:18:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' b030130f-934a-44e3-894e-27483fe25d3f '!=' b030130f-934a-44e3-894e-27483fe25d3f ']' 00:12:13.970 15:18:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 85314 00:12:13.970 15:18:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 85314 ']' 00:12:13.970 15:18:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 85314 00:12:13.970 15:18:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:12:13.970 15:18:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:12:13.970 15:18:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 85314 00:12:14.230 15:18:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:12:14.230 15:18:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:12:14.230 killing process with pid 85314 00:12:14.230 15:18:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 85314' 00:12:14.230 15:18:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 85314 00:12:14.230 [2024-11-19 15:18:04.316570] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:14.230 15:18:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 85314 00:12:14.230 [2024-11-19 15:18:04.316712] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:14.230 [2024-11-19 15:18:04.316806] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:14.230 [2024-11-19 15:18:04.316825] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:12:14.230 [2024-11-19 15:18:04.401012] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:14.490 15:18:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:12:14.490 00:12:14.490 real 0m7.052s 00:12:14.490 user 0m11.741s 00:12:14.490 sys 0m1.498s 00:12:14.490 15:18:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:12:14.490 15:18:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:12:14.490 ************************************ 00:12:14.490 END TEST raid_superblock_test 00:12:14.490 ************************************ 00:12:14.490 15:18:04 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid1 4 read 00:12:14.490 15:18:04 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:12:14.490 15:18:04 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:12:14.490 15:18:04 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:14.490 ************************************ 00:12:14.490 START TEST raid_read_error_test 00:12:14.490 ************************************ 00:12:14.490 15:18:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid1 4 read 00:12:14.490 15:18:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:12:14.490 15:18:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:12:14.490 15:18:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:12:14.490 15:18:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:12:14.490 15:18:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:12:14.490 15:18:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:12:14.490 15:18:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:12:14.490 15:18:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:12:14.490 15:18:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:12:14.490 15:18:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:12:14.490 15:18:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:12:14.490 15:18:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:12:14.490 15:18:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:12:14.490 15:18:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:12:14.490 15:18:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:12:14.490 15:18:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:12:14.490 15:18:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:12:14.490 15:18:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:12:14.490 15:18:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:12:14.490 15:18:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:12:14.490 15:18:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:12:14.490 15:18:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:12:14.490 15:18:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:12:14.490 15:18:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:12:14.490 15:18:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:12:14.490 15:18:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:12:14.490 15:18:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:12:14.490 15:18:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.hWZWt7d8mx 00:12:14.490 15:18:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=85785 00:12:14.490 15:18:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:12:14.490 15:18:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 85785 00:12:14.490 15:18:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 85785 ']' 00:12:14.490 15:18:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:14.490 15:18:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:12:14.490 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:14.490 15:18:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:14.491 15:18:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:12:14.491 15:18:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:14.753 [2024-11-19 15:18:04.896319] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:12:14.753 [2024-11-19 15:18:04.896475] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid85785 ] 00:12:14.753 [2024-11-19 15:18:05.054588] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:15.019 [2024-11-19 15:18:05.096050] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:12:15.019 [2024-11-19 15:18:05.172653] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:15.019 [2024-11-19 15:18:05.172698] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:15.597 BaseBdev1_malloc 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:15.597 true 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:15.597 [2024-11-19 15:18:05.755933] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:12:15.597 [2024-11-19 15:18:05.756028] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:15.597 [2024-11-19 15:18:05.756057] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:12:15.597 [2024-11-19 15:18:05.756078] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:15.597 [2024-11-19 15:18:05.758686] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:15.597 [2024-11-19 15:18:05.758722] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:15.597 BaseBdev1 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:15.597 BaseBdev2_malloc 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:15.597 true 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:15.597 [2024-11-19 15:18:05.803056] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:12:15.597 [2024-11-19 15:18:05.803135] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:15.597 [2024-11-19 15:18:05.803162] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:12:15.597 [2024-11-19 15:18:05.803183] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:15.597 [2024-11-19 15:18:05.805811] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:15.597 [2024-11-19 15:18:05.805856] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:12:15.597 BaseBdev2 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:15.597 BaseBdev3_malloc 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:15.597 true 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:15.597 [2024-11-19 15:18:05.850129] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:12:15.597 [2024-11-19 15:18:05.850193] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:15.597 [2024-11-19 15:18:05.850212] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:12:15.597 [2024-11-19 15:18:05.850222] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:15.597 [2024-11-19 15:18:05.852522] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:15.597 [2024-11-19 15:18:05.852559] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:12:15.597 BaseBdev3 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:15.597 BaseBdev4_malloc 00:12:15.597 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:15.598 15:18:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:12:15.598 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:15.598 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:15.598 true 00:12:15.598 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:15.598 15:18:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:12:15.598 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:15.598 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:15.598 [2024-11-19 15:18:05.905577] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:12:15.598 [2024-11-19 15:18:05.905644] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:15.598 [2024-11-19 15:18:05.905671] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:12:15.598 [2024-11-19 15:18:05.905678] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:15.598 [2024-11-19 15:18:05.907938] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:15.598 [2024-11-19 15:18:05.907983] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:12:15.598 BaseBdev4 00:12:15.598 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:15.598 15:18:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:12:15.598 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:15.598 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:15.598 [2024-11-19 15:18:05.917610] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:15.598 [2024-11-19 15:18:05.919720] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:15.598 [2024-11-19 15:18:05.919801] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:15.598 [2024-11-19 15:18:05.919861] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:12:15.598 [2024-11-19 15:18:05.920085] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002000 00:12:15.598 [2024-11-19 15:18:05.920111] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:15.598 [2024-11-19 15:18:05.920394] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002ef0 00:12:15.598 [2024-11-19 15:18:05.920560] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002000 00:12:15.598 [2024-11-19 15:18:05.920580] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002000 00:12:15.598 [2024-11-19 15:18:05.920718] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:15.598 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:15.598 15:18:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:12:15.598 15:18:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:15.598 15:18:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:15.598 15:18:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:15.598 15:18:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:15.598 15:18:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:15.598 15:18:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:15.598 15:18:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:15.598 15:18:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:15.598 15:18:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:15.598 15:18:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:15.598 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:15.598 15:18:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:15.598 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:15.858 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:15.858 15:18:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:15.858 "name": "raid_bdev1", 00:12:15.858 "uuid": "c668b956-79c1-48ae-980f-edc6e1279a29", 00:12:15.858 "strip_size_kb": 0, 00:12:15.858 "state": "online", 00:12:15.858 "raid_level": "raid1", 00:12:15.858 "superblock": true, 00:12:15.858 "num_base_bdevs": 4, 00:12:15.858 "num_base_bdevs_discovered": 4, 00:12:15.858 "num_base_bdevs_operational": 4, 00:12:15.858 "base_bdevs_list": [ 00:12:15.858 { 00:12:15.858 "name": "BaseBdev1", 00:12:15.858 "uuid": "467cbc19-203a-506d-96f4-e799fa9ab66c", 00:12:15.858 "is_configured": true, 00:12:15.858 "data_offset": 2048, 00:12:15.858 "data_size": 63488 00:12:15.858 }, 00:12:15.858 { 00:12:15.858 "name": "BaseBdev2", 00:12:15.858 "uuid": "f5ae4980-456a-551c-bef7-d3da26af4cb3", 00:12:15.858 "is_configured": true, 00:12:15.858 "data_offset": 2048, 00:12:15.858 "data_size": 63488 00:12:15.858 }, 00:12:15.858 { 00:12:15.858 "name": "BaseBdev3", 00:12:15.858 "uuid": "4434783b-75a5-54ad-ac8a-543a525172ec", 00:12:15.858 "is_configured": true, 00:12:15.858 "data_offset": 2048, 00:12:15.858 "data_size": 63488 00:12:15.858 }, 00:12:15.858 { 00:12:15.858 "name": "BaseBdev4", 00:12:15.858 "uuid": "bad1bd51-9245-5b92-a5b0-0219a6c003e2", 00:12:15.858 "is_configured": true, 00:12:15.858 "data_offset": 2048, 00:12:15.858 "data_size": 63488 00:12:15.858 } 00:12:15.858 ] 00:12:15.858 }' 00:12:15.858 15:18:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:15.858 15:18:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:16.118 15:18:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:12:16.118 15:18:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:12:16.118 [2024-11-19 15:18:06.449259] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000003090 00:12:17.058 15:18:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:12:17.058 15:18:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:17.058 15:18:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:17.058 15:18:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:17.058 15:18:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:12:17.058 15:18:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:12:17.058 15:18:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ read = \w\r\i\t\e ]] 00:12:17.058 15:18:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:12:17.058 15:18:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:12:17.058 15:18:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:17.058 15:18:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:17.058 15:18:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:17.058 15:18:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:17.058 15:18:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:17.058 15:18:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:17.058 15:18:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:17.058 15:18:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:17.058 15:18:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:17.058 15:18:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:17.058 15:18:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:17.058 15:18:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:17.058 15:18:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:17.058 15:18:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:17.318 15:18:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:17.318 "name": "raid_bdev1", 00:12:17.318 "uuid": "c668b956-79c1-48ae-980f-edc6e1279a29", 00:12:17.318 "strip_size_kb": 0, 00:12:17.318 "state": "online", 00:12:17.318 "raid_level": "raid1", 00:12:17.318 "superblock": true, 00:12:17.318 "num_base_bdevs": 4, 00:12:17.318 "num_base_bdevs_discovered": 4, 00:12:17.318 "num_base_bdevs_operational": 4, 00:12:17.318 "base_bdevs_list": [ 00:12:17.318 { 00:12:17.318 "name": "BaseBdev1", 00:12:17.318 "uuid": "467cbc19-203a-506d-96f4-e799fa9ab66c", 00:12:17.318 "is_configured": true, 00:12:17.318 "data_offset": 2048, 00:12:17.318 "data_size": 63488 00:12:17.318 }, 00:12:17.318 { 00:12:17.318 "name": "BaseBdev2", 00:12:17.318 "uuid": "f5ae4980-456a-551c-bef7-d3da26af4cb3", 00:12:17.318 "is_configured": true, 00:12:17.318 "data_offset": 2048, 00:12:17.318 "data_size": 63488 00:12:17.318 }, 00:12:17.318 { 00:12:17.318 "name": "BaseBdev3", 00:12:17.318 "uuid": "4434783b-75a5-54ad-ac8a-543a525172ec", 00:12:17.318 "is_configured": true, 00:12:17.318 "data_offset": 2048, 00:12:17.318 "data_size": 63488 00:12:17.318 }, 00:12:17.318 { 00:12:17.318 "name": "BaseBdev4", 00:12:17.318 "uuid": "bad1bd51-9245-5b92-a5b0-0219a6c003e2", 00:12:17.318 "is_configured": true, 00:12:17.318 "data_offset": 2048, 00:12:17.318 "data_size": 63488 00:12:17.318 } 00:12:17.318 ] 00:12:17.318 }' 00:12:17.318 15:18:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:17.318 15:18:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:17.578 15:18:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:12:17.578 15:18:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:17.578 15:18:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:17.578 [2024-11-19 15:18:07.809709] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:17.578 [2024-11-19 15:18:07.809762] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:17.578 [2024-11-19 15:18:07.812192] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:17.578 [2024-11-19 15:18:07.812250] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:17.578 [2024-11-19 15:18:07.812395] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:17.578 [2024-11-19 15:18:07.812423] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state offline 00:12:17.578 { 00:12:17.578 "results": [ 00:12:17.578 { 00:12:17.578 "job": "raid_bdev1", 00:12:17.578 "core_mask": "0x1", 00:12:17.578 "workload": "randrw", 00:12:17.578 "percentage": 50, 00:12:17.578 "status": "finished", 00:12:17.578 "queue_depth": 1, 00:12:17.578 "io_size": 131072, 00:12:17.578 "runtime": 1.360924, 00:12:17.578 "iops": 8325.226096387454, 00:12:17.578 "mibps": 1040.6532620484318, 00:12:17.578 "io_failed": 0, 00:12:17.578 "io_timeout": 0, 00:12:17.578 "avg_latency_us": 117.54858492929463, 00:12:17.578 "min_latency_us": 22.581659388646287, 00:12:17.578 "max_latency_us": 1495.3082969432314 00:12:17.578 } 00:12:17.578 ], 00:12:17.578 "core_count": 1 00:12:17.578 } 00:12:17.578 15:18:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:17.578 15:18:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 85785 00:12:17.578 15:18:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 85785 ']' 00:12:17.578 15:18:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 85785 00:12:17.578 15:18:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:12:17.578 15:18:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:12:17.579 15:18:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 85785 00:12:17.579 15:18:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:12:17.579 15:18:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:12:17.579 killing process with pid 85785 00:12:17.579 15:18:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 85785' 00:12:17.579 15:18:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 85785 00:12:17.579 [2024-11-19 15:18:07.860647] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:17.579 15:18:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 85785 00:12:17.838 [2024-11-19 15:18:07.931000] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:18.099 15:18:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.hWZWt7d8mx 00:12:18.099 15:18:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:12:18.099 15:18:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:12:18.099 15:18:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:12:18.099 15:18:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:12:18.099 15:18:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:12:18.099 15:18:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:12:18.099 15:18:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:12:18.099 00:12:18.099 real 0m3.471s 00:12:18.099 user 0m4.245s 00:12:18.099 sys 0m0.648s 00:12:18.099 15:18:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:12:18.099 15:18:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:18.099 ************************************ 00:12:18.099 END TEST raid_read_error_test 00:12:18.099 ************************************ 00:12:18.099 15:18:08 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid1 4 write 00:12:18.099 15:18:08 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:12:18.099 15:18:08 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:12:18.099 15:18:08 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:18.099 ************************************ 00:12:18.099 START TEST raid_write_error_test 00:12:18.099 ************************************ 00:12:18.099 15:18:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid1 4 write 00:12:18.099 15:18:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:12:18.099 15:18:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:12:18.099 15:18:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:12:18.099 15:18:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:12:18.099 15:18:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:12:18.099 15:18:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:12:18.099 15:18:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:12:18.099 15:18:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:12:18.099 15:18:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:12:18.099 15:18:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:12:18.099 15:18:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:12:18.099 15:18:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:12:18.099 15:18:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:12:18.099 15:18:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:12:18.099 15:18:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:12:18.099 15:18:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:12:18.099 15:18:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:12:18.099 15:18:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:12:18.099 15:18:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:12:18.099 15:18:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:12:18.099 15:18:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:12:18.099 15:18:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:12:18.099 15:18:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:12:18.099 15:18:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:12:18.099 15:18:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:12:18.099 15:18:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:12:18.099 15:18:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:12:18.099 15:18:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.xtdpHRxXvI 00:12:18.099 15:18:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=85914 00:12:18.099 15:18:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:12:18.099 15:18:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 85914 00:12:18.099 15:18:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 85914 ']' 00:12:18.099 15:18:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:18.099 15:18:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:12:18.099 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:18.099 15:18:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:18.099 15:18:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:12:18.099 15:18:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:18.099 [2024-11-19 15:18:08.433546] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:12:18.099 [2024-11-19 15:18:08.433707] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid85914 ] 00:12:18.359 [2024-11-19 15:18:08.587587] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:18.359 [2024-11-19 15:18:08.631159] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:12:18.619 [2024-11-19 15:18:08.707658] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:18.619 [2024-11-19 15:18:08.707718] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:19.189 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:12:19.189 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:12:19.189 15:18:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:12:19.189 15:18:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:12:19.189 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:19.189 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:19.189 BaseBdev1_malloc 00:12:19.189 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:19.189 15:18:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:12:19.189 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:19.189 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:19.189 true 00:12:19.189 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:19.189 15:18:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:12:19.189 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:19.189 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:19.189 [2024-11-19 15:18:09.285564] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:12:19.189 [2024-11-19 15:18:09.285640] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:19.189 [2024-11-19 15:18:09.285666] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:12:19.189 [2024-11-19 15:18:09.285675] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:19.189 [2024-11-19 15:18:09.288173] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:19.189 [2024-11-19 15:18:09.288207] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:19.189 BaseBdev1 00:12:19.189 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:19.189 15:18:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:12:19.189 15:18:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:12:19.189 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:19.189 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:19.189 BaseBdev2_malloc 00:12:19.189 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:19.189 15:18:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:12:19.189 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:19.189 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:19.189 true 00:12:19.189 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:19.189 15:18:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:12:19.189 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:19.189 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:19.189 [2024-11-19 15:18:09.332085] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:12:19.189 [2024-11-19 15:18:09.332137] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:19.190 [2024-11-19 15:18:09.332156] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:12:19.190 [2024-11-19 15:18:09.332175] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:19.190 [2024-11-19 15:18:09.334508] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:19.190 [2024-11-19 15:18:09.334563] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:12:19.190 BaseBdev2 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:19.190 BaseBdev3_malloc 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:19.190 true 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:19.190 [2024-11-19 15:18:09.378567] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:12:19.190 [2024-11-19 15:18:09.378616] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:19.190 [2024-11-19 15:18:09.378636] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:12:19.190 [2024-11-19 15:18:09.378644] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:19.190 [2024-11-19 15:18:09.381142] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:19.190 [2024-11-19 15:18:09.381175] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:12:19.190 BaseBdev3 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:19.190 BaseBdev4_malloc 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:19.190 true 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:19.190 [2024-11-19 15:18:09.434456] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:12:19.190 [2024-11-19 15:18:09.434517] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:19.190 [2024-11-19 15:18:09.434543] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:12:19.190 [2024-11-19 15:18:09.434552] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:19.190 [2024-11-19 15:18:09.436921] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:19.190 [2024-11-19 15:18:09.436958] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:12:19.190 BaseBdev4 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:19.190 [2024-11-19 15:18:09.446503] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:19.190 [2024-11-19 15:18:09.448717] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:19.190 [2024-11-19 15:18:09.448800] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:19.190 [2024-11-19 15:18:09.448862] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:12:19.190 [2024-11-19 15:18:09.449119] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002000 00:12:19.190 [2024-11-19 15:18:09.449139] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:19.190 [2024-11-19 15:18:09.449435] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002ef0 00:12:19.190 [2024-11-19 15:18:09.449594] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002000 00:12:19.190 [2024-11-19 15:18:09.449623] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002000 00:12:19.190 [2024-11-19 15:18:09.449767] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:19.190 "name": "raid_bdev1", 00:12:19.190 "uuid": "866221b6-4a11-4522-96ed-180e554279c2", 00:12:19.190 "strip_size_kb": 0, 00:12:19.190 "state": "online", 00:12:19.190 "raid_level": "raid1", 00:12:19.190 "superblock": true, 00:12:19.190 "num_base_bdevs": 4, 00:12:19.190 "num_base_bdevs_discovered": 4, 00:12:19.190 "num_base_bdevs_operational": 4, 00:12:19.190 "base_bdevs_list": [ 00:12:19.190 { 00:12:19.190 "name": "BaseBdev1", 00:12:19.190 "uuid": "5a8dff68-4b31-5d29-8ab7-d1e50d803bd1", 00:12:19.190 "is_configured": true, 00:12:19.190 "data_offset": 2048, 00:12:19.190 "data_size": 63488 00:12:19.190 }, 00:12:19.190 { 00:12:19.190 "name": "BaseBdev2", 00:12:19.190 "uuid": "c4c9dff0-dfc2-5002-bcc8-4a829312cecc", 00:12:19.190 "is_configured": true, 00:12:19.190 "data_offset": 2048, 00:12:19.190 "data_size": 63488 00:12:19.190 }, 00:12:19.190 { 00:12:19.190 "name": "BaseBdev3", 00:12:19.190 "uuid": "2169c50f-a69a-5157-b342-2a81e56e7c46", 00:12:19.190 "is_configured": true, 00:12:19.190 "data_offset": 2048, 00:12:19.190 "data_size": 63488 00:12:19.190 }, 00:12:19.190 { 00:12:19.190 "name": "BaseBdev4", 00:12:19.190 "uuid": "1676683a-358a-5774-91ed-140c21530981", 00:12:19.190 "is_configured": true, 00:12:19.190 "data_offset": 2048, 00:12:19.190 "data_size": 63488 00:12:19.190 } 00:12:19.190 ] 00:12:19.190 }' 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:19.190 15:18:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:19.761 15:18:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:12:19.761 15:18:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:12:19.761 [2024-11-19 15:18:09.982175] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000003090 00:12:20.701 15:18:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:12:20.701 15:18:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:20.701 15:18:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:20.701 [2024-11-19 15:18:10.899811] bdev_raid.c:2276:_raid_bdev_fail_base_bdev: *NOTICE*: Failing base bdev in slot 0 ('BaseBdev1') of raid bdev 'raid_bdev1' 00:12:20.701 [2024-11-19 15:18:10.899884] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:12:20.701 [2024-11-19 15:18:10.900160] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000003090 00:12:20.701 15:18:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:20.701 15:18:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:12:20.701 15:18:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:12:20.701 15:18:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ write = \w\r\i\t\e ]] 00:12:20.701 15:18:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=3 00:12:20.701 15:18:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:20.701 15:18:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:20.701 15:18:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:20.701 15:18:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:20.701 15:18:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:20.701 15:18:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:20.701 15:18:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:20.701 15:18:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:20.701 15:18:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:20.701 15:18:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:20.701 15:18:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:20.701 15:18:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:20.701 15:18:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:20.701 15:18:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:20.701 15:18:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:20.701 15:18:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:20.701 "name": "raid_bdev1", 00:12:20.701 "uuid": "866221b6-4a11-4522-96ed-180e554279c2", 00:12:20.701 "strip_size_kb": 0, 00:12:20.701 "state": "online", 00:12:20.701 "raid_level": "raid1", 00:12:20.701 "superblock": true, 00:12:20.701 "num_base_bdevs": 4, 00:12:20.701 "num_base_bdevs_discovered": 3, 00:12:20.701 "num_base_bdevs_operational": 3, 00:12:20.701 "base_bdevs_list": [ 00:12:20.701 { 00:12:20.701 "name": null, 00:12:20.701 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:20.701 "is_configured": false, 00:12:20.701 "data_offset": 0, 00:12:20.701 "data_size": 63488 00:12:20.701 }, 00:12:20.701 { 00:12:20.701 "name": "BaseBdev2", 00:12:20.701 "uuid": "c4c9dff0-dfc2-5002-bcc8-4a829312cecc", 00:12:20.701 "is_configured": true, 00:12:20.701 "data_offset": 2048, 00:12:20.701 "data_size": 63488 00:12:20.701 }, 00:12:20.701 { 00:12:20.701 "name": "BaseBdev3", 00:12:20.701 "uuid": "2169c50f-a69a-5157-b342-2a81e56e7c46", 00:12:20.701 "is_configured": true, 00:12:20.701 "data_offset": 2048, 00:12:20.701 "data_size": 63488 00:12:20.701 }, 00:12:20.701 { 00:12:20.701 "name": "BaseBdev4", 00:12:20.701 "uuid": "1676683a-358a-5774-91ed-140c21530981", 00:12:20.701 "is_configured": true, 00:12:20.701 "data_offset": 2048, 00:12:20.701 "data_size": 63488 00:12:20.701 } 00:12:20.701 ] 00:12:20.701 }' 00:12:20.701 15:18:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:20.701 15:18:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:21.270 15:18:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:12:21.270 15:18:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:21.270 15:18:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:21.270 [2024-11-19 15:18:11.344506] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:21.270 [2024-11-19 15:18:11.344565] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:21.270 [2024-11-19 15:18:11.347029] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:21.270 [2024-11-19 15:18:11.347101] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:21.270 [2024-11-19 15:18:11.347212] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:21.270 [2024-11-19 15:18:11.347225] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state offline 00:12:21.270 { 00:12:21.270 "results": [ 00:12:21.270 { 00:12:21.270 "job": "raid_bdev1", 00:12:21.270 "core_mask": "0x1", 00:12:21.270 "workload": "randrw", 00:12:21.270 "percentage": 50, 00:12:21.270 "status": "finished", 00:12:21.270 "queue_depth": 1, 00:12:21.270 "io_size": 131072, 00:12:21.270 "runtime": 1.362737, 00:12:21.270 "iops": 8882.858541303274, 00:12:21.270 "mibps": 1110.3573176629093, 00:12:21.270 "io_failed": 0, 00:12:21.270 "io_timeout": 0, 00:12:21.270 "avg_latency_us": 109.9396917438209, 00:12:21.270 "min_latency_us": 22.246288209606988, 00:12:21.270 "max_latency_us": 1602.6270742358079 00:12:21.270 } 00:12:21.270 ], 00:12:21.270 "core_count": 1 00:12:21.270 } 00:12:21.270 15:18:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:21.270 15:18:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 85914 00:12:21.270 15:18:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 85914 ']' 00:12:21.270 15:18:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 85914 00:12:21.270 15:18:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:12:21.270 15:18:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:12:21.270 15:18:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 85914 00:12:21.270 15:18:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:12:21.270 15:18:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:12:21.270 killing process with pid 85914 00:12:21.270 15:18:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 85914' 00:12:21.270 15:18:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 85914 00:12:21.270 [2024-11-19 15:18:11.389658] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:21.270 15:18:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 85914 00:12:21.270 [2024-11-19 15:18:11.458809] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:21.530 15:18:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.xtdpHRxXvI 00:12:21.530 15:18:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:12:21.530 15:18:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:12:21.530 15:18:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:12:21.530 15:18:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:12:21.530 15:18:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:12:21.530 15:18:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:12:21.530 15:18:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:12:21.530 00:12:21.530 real 0m3.451s 00:12:21.530 user 0m4.233s 00:12:21.530 sys 0m0.629s 00:12:21.530 15:18:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:12:21.530 15:18:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:12:21.530 ************************************ 00:12:21.530 END TEST raid_write_error_test 00:12:21.530 ************************************ 00:12:21.530 15:18:11 bdev_raid -- bdev/bdev_raid.sh@976 -- # '[' true = true ']' 00:12:21.530 15:18:11 bdev_raid -- bdev/bdev_raid.sh@977 -- # for n in 2 4 00:12:21.530 15:18:11 bdev_raid -- bdev/bdev_raid.sh@978 -- # run_test raid_rebuild_test raid_rebuild_test raid1 2 false false true 00:12:21.530 15:18:11 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:12:21.530 15:18:11 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:12:21.530 15:18:11 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:21.530 ************************************ 00:12:21.530 START TEST raid_rebuild_test 00:12:21.530 ************************************ 00:12:21.530 15:18:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 false false true 00:12:21.530 15:18:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:12:21.530 15:18:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:12:21.531 15:18:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:12:21.531 15:18:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:12:21.531 15:18:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:12:21.531 15:18:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:12:21.531 15:18:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:21.531 15:18:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:12:21.531 15:18:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:21.531 15:18:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:21.531 15:18:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:12:21.531 15:18:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:21.531 15:18:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:21.531 15:18:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:12:21.531 15:18:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:12:21.531 15:18:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:12:21.531 15:18:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:12:21.531 15:18:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:12:21.531 15:18:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:12:21.531 15:18:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:12:21.531 15:18:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:12:21.531 15:18:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:12:21.531 15:18:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:12:21.531 15:18:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=86047 00:12:21.531 15:18:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 86047 00:12:21.531 15:18:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:12:21.531 15:18:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@835 -- # '[' -z 86047 ']' 00:12:21.531 15:18:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:21.531 15:18:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:12:21.531 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:21.531 15:18:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:21.531 15:18:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:12:21.531 15:18:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:21.791 I/O size of 3145728 is greater than zero copy threshold (65536). 00:12:21.791 Zero copy mechanism will not be used. 00:12:21.791 [2024-11-19 15:18:11.938663] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:12:21.791 [2024-11-19 15:18:11.938795] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid86047 ] 00:12:21.791 [2024-11-19 15:18:12.075410] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:21.791 [2024-11-19 15:18:12.116393] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:12:22.055 [2024-11-19 15:18:12.192782] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:22.055 [2024-11-19 15:18:12.192840] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:22.661 15:18:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:12:22.661 15:18:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@868 -- # return 0 00:12:22.661 15:18:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:22.661 15:18:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:12:22.661 15:18:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:22.661 15:18:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:22.661 BaseBdev1_malloc 00:12:22.661 15:18:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:22.661 15:18:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:22.661 15:18:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:22.661 15:18:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:22.661 [2024-11-19 15:18:12.830824] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:22.661 [2024-11-19 15:18:12.830902] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:22.661 [2024-11-19 15:18:12.830936] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:12:22.661 [2024-11-19 15:18:12.830950] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:22.661 [2024-11-19 15:18:12.833625] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:22.661 [2024-11-19 15:18:12.833666] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:22.661 BaseBdev1 00:12:22.661 15:18:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:22.661 15:18:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:22.661 15:18:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:12:22.661 15:18:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:22.661 15:18:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:22.661 BaseBdev2_malloc 00:12:22.661 15:18:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:22.661 15:18:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:12:22.661 15:18:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:22.662 15:18:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:22.662 [2024-11-19 15:18:12.865794] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:12:22.662 [2024-11-19 15:18:12.865852] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:22.662 [2024-11-19 15:18:12.865874] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:12:22.662 [2024-11-19 15:18:12.865884] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:22.662 [2024-11-19 15:18:12.868476] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:22.662 [2024-11-19 15:18:12.868519] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:12:22.662 BaseBdev2 00:12:22.662 15:18:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:22.662 15:18:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:12:22.662 15:18:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:22.662 15:18:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:22.662 spare_malloc 00:12:22.662 15:18:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:22.662 15:18:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:12:22.662 15:18:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:22.662 15:18:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:22.662 spare_delay 00:12:22.662 15:18:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:22.662 15:18:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:22.662 15:18:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:22.662 15:18:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:22.662 [2024-11-19 15:18:12.908661] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:22.662 [2024-11-19 15:18:12.908728] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:22.662 [2024-11-19 15:18:12.908752] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:12:22.662 [2024-11-19 15:18:12.908762] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:22.662 [2024-11-19 15:18:12.911201] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:22.662 [2024-11-19 15:18:12.911236] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:22.662 spare 00:12:22.662 15:18:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:22.662 15:18:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:12:22.662 15:18:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:22.662 15:18:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:22.662 [2024-11-19 15:18:12.916695] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:22.662 [2024-11-19 15:18:12.918817] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:22.662 [2024-11-19 15:18:12.918913] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:12:22.662 [2024-11-19 15:18:12.918931] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:12:22.662 [2024-11-19 15:18:12.919237] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:12:22.662 [2024-11-19 15:18:12.919396] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:12:22.662 [2024-11-19 15:18:12.919418] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:12:22.662 [2024-11-19 15:18:12.919533] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:22.662 15:18:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:22.662 15:18:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:22.662 15:18:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:22.662 15:18:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:22.662 15:18:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:22.662 15:18:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:22.662 15:18:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:22.662 15:18:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:22.662 15:18:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:22.662 15:18:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:22.662 15:18:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:22.662 15:18:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:22.662 15:18:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:22.662 15:18:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:22.662 15:18:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:22.662 15:18:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:22.662 15:18:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:22.662 "name": "raid_bdev1", 00:12:22.662 "uuid": "b40e3c58-b86f-404e-a0c1-407da15de86a", 00:12:22.662 "strip_size_kb": 0, 00:12:22.662 "state": "online", 00:12:22.662 "raid_level": "raid1", 00:12:22.662 "superblock": false, 00:12:22.662 "num_base_bdevs": 2, 00:12:22.662 "num_base_bdevs_discovered": 2, 00:12:22.662 "num_base_bdevs_operational": 2, 00:12:22.662 "base_bdevs_list": [ 00:12:22.662 { 00:12:22.662 "name": "BaseBdev1", 00:12:22.662 "uuid": "42aedd7c-42ce-5d18-ac6e-cec064064d76", 00:12:22.662 "is_configured": true, 00:12:22.662 "data_offset": 0, 00:12:22.662 "data_size": 65536 00:12:22.662 }, 00:12:22.662 { 00:12:22.662 "name": "BaseBdev2", 00:12:22.662 "uuid": "3315fdd4-a5c2-51c0-a37b-4171bac80265", 00:12:22.662 "is_configured": true, 00:12:22.662 "data_offset": 0, 00:12:22.662 "data_size": 65536 00:12:22.662 } 00:12:22.662 ] 00:12:22.662 }' 00:12:22.662 15:18:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:22.662 15:18:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:23.232 15:18:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:23.232 15:18:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:23.232 15:18:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:23.232 15:18:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:12:23.232 [2024-11-19 15:18:13.352340] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:23.232 15:18:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:23.232 15:18:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:12:23.232 15:18:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:23.232 15:18:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:12:23.232 15:18:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:23.232 15:18:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:23.232 15:18:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:23.232 15:18:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:12:23.232 15:18:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:12:23.232 15:18:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:12:23.232 15:18:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:12:23.232 15:18:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:12:23.232 15:18:13 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:23.232 15:18:13 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:12:23.232 15:18:13 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:23.232 15:18:13 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:12:23.232 15:18:13 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:23.232 15:18:13 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:12:23.232 15:18:13 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:23.232 15:18:13 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:23.232 15:18:13 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:12:23.492 [2024-11-19 15:18:13.631973] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:12:23.492 /dev/nbd0 00:12:23.492 15:18:13 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:23.492 15:18:13 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:23.492 15:18:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:12:23.492 15:18:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:12:23.492 15:18:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:23.492 15:18:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:23.492 15:18:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:12:23.492 15:18:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@877 -- # break 00:12:23.492 15:18:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:23.492 15:18:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:23.492 15:18:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:23.492 1+0 records in 00:12:23.492 1+0 records out 00:12:23.492 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000474898 s, 8.6 MB/s 00:12:23.492 15:18:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:23.492 15:18:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:12:23.492 15:18:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:23.492 15:18:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:23.492 15:18:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:12:23.492 15:18:13 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:23.492 15:18:13 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:23.492 15:18:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:12:23.492 15:18:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:12:23.492 15:18:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=65536 oflag=direct 00:12:28.769 65536+0 records in 00:12:28.769 65536+0 records out 00:12:28.769 33554432 bytes (34 MB, 32 MiB) copied, 4.46879 s, 7.5 MB/s 00:12:28.769 15:18:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:12:28.769 15:18:18 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:28.769 15:18:18 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:12:28.769 15:18:18 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:28.769 15:18:18 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:12:28.769 15:18:18 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:28.769 15:18:18 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:28.769 [2024-11-19 15:18:18.360625] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:28.769 15:18:18 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:28.769 15:18:18 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:28.769 15:18:18 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:28.769 15:18:18 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:28.769 15:18:18 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:28.769 15:18:18 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:28.769 15:18:18 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:12:28.769 15:18:18 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:12:28.769 15:18:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:12:28.769 15:18:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:28.769 15:18:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:28.769 [2024-11-19 15:18:18.389991] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:12:28.769 15:18:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:28.769 15:18:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:28.769 15:18:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:28.769 15:18:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:28.769 15:18:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:28.769 15:18:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:28.769 15:18:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:28.769 15:18:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:28.769 15:18:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:28.769 15:18:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:28.769 15:18:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:28.769 15:18:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:28.769 15:18:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:28.769 15:18:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:28.769 15:18:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:28.769 15:18:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:28.769 15:18:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:28.769 "name": "raid_bdev1", 00:12:28.769 "uuid": "b40e3c58-b86f-404e-a0c1-407da15de86a", 00:12:28.769 "strip_size_kb": 0, 00:12:28.769 "state": "online", 00:12:28.769 "raid_level": "raid1", 00:12:28.769 "superblock": false, 00:12:28.769 "num_base_bdevs": 2, 00:12:28.769 "num_base_bdevs_discovered": 1, 00:12:28.769 "num_base_bdevs_operational": 1, 00:12:28.769 "base_bdevs_list": [ 00:12:28.769 { 00:12:28.769 "name": null, 00:12:28.769 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:28.769 "is_configured": false, 00:12:28.769 "data_offset": 0, 00:12:28.769 "data_size": 65536 00:12:28.769 }, 00:12:28.769 { 00:12:28.769 "name": "BaseBdev2", 00:12:28.769 "uuid": "3315fdd4-a5c2-51c0-a37b-4171bac80265", 00:12:28.769 "is_configured": true, 00:12:28.769 "data_offset": 0, 00:12:28.769 "data_size": 65536 00:12:28.769 } 00:12:28.769 ] 00:12:28.769 }' 00:12:28.769 15:18:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:28.769 15:18:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:28.769 15:18:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:28.769 15:18:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:28.769 15:18:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:28.769 [2024-11-19 15:18:18.789297] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:28.769 [2024-11-19 15:18:18.809193] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d06220 00:12:28.769 15:18:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:28.769 15:18:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:12:28.769 [2024-11-19 15:18:18.812068] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:29.718 15:18:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:29.718 15:18:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:29.718 15:18:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:29.718 15:18:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:29.718 15:18:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:29.718 15:18:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:29.718 15:18:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:29.718 15:18:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:29.718 15:18:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:29.718 15:18:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:29.718 15:18:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:29.718 "name": "raid_bdev1", 00:12:29.718 "uuid": "b40e3c58-b86f-404e-a0c1-407da15de86a", 00:12:29.718 "strip_size_kb": 0, 00:12:29.718 "state": "online", 00:12:29.718 "raid_level": "raid1", 00:12:29.718 "superblock": false, 00:12:29.718 "num_base_bdevs": 2, 00:12:29.718 "num_base_bdevs_discovered": 2, 00:12:29.718 "num_base_bdevs_operational": 2, 00:12:29.718 "process": { 00:12:29.718 "type": "rebuild", 00:12:29.718 "target": "spare", 00:12:29.718 "progress": { 00:12:29.718 "blocks": 20480, 00:12:29.718 "percent": 31 00:12:29.718 } 00:12:29.718 }, 00:12:29.718 "base_bdevs_list": [ 00:12:29.718 { 00:12:29.718 "name": "spare", 00:12:29.718 "uuid": "05ba46d7-1719-55d6-8164-42e5b53afd28", 00:12:29.718 "is_configured": true, 00:12:29.718 "data_offset": 0, 00:12:29.718 "data_size": 65536 00:12:29.718 }, 00:12:29.718 { 00:12:29.718 "name": "BaseBdev2", 00:12:29.718 "uuid": "3315fdd4-a5c2-51c0-a37b-4171bac80265", 00:12:29.718 "is_configured": true, 00:12:29.718 "data_offset": 0, 00:12:29.718 "data_size": 65536 00:12:29.718 } 00:12:29.718 ] 00:12:29.718 }' 00:12:29.718 15:18:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:29.718 15:18:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:29.718 15:18:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:29.718 15:18:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:29.718 15:18:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:29.718 15:18:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:29.718 15:18:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:29.718 [2024-11-19 15:18:19.952390] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:29.718 [2024-11-19 15:18:20.022142] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:29.718 [2024-11-19 15:18:20.022228] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:29.718 [2024-11-19 15:18:20.022252] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:29.718 [2024-11-19 15:18:20.022259] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:29.718 15:18:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:29.718 15:18:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:29.718 15:18:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:29.718 15:18:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:29.718 15:18:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:29.718 15:18:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:29.718 15:18:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:29.718 15:18:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:29.718 15:18:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:29.718 15:18:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:29.718 15:18:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:29.718 15:18:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:29.718 15:18:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:29.718 15:18:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:29.718 15:18:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:29.978 15:18:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:29.978 15:18:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:29.978 "name": "raid_bdev1", 00:12:29.978 "uuid": "b40e3c58-b86f-404e-a0c1-407da15de86a", 00:12:29.978 "strip_size_kb": 0, 00:12:29.978 "state": "online", 00:12:29.978 "raid_level": "raid1", 00:12:29.978 "superblock": false, 00:12:29.978 "num_base_bdevs": 2, 00:12:29.978 "num_base_bdevs_discovered": 1, 00:12:29.978 "num_base_bdevs_operational": 1, 00:12:29.978 "base_bdevs_list": [ 00:12:29.978 { 00:12:29.978 "name": null, 00:12:29.978 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:29.978 "is_configured": false, 00:12:29.978 "data_offset": 0, 00:12:29.978 "data_size": 65536 00:12:29.978 }, 00:12:29.978 { 00:12:29.978 "name": "BaseBdev2", 00:12:29.978 "uuid": "3315fdd4-a5c2-51c0-a37b-4171bac80265", 00:12:29.978 "is_configured": true, 00:12:29.978 "data_offset": 0, 00:12:29.978 "data_size": 65536 00:12:29.978 } 00:12:29.978 ] 00:12:29.978 }' 00:12:29.978 15:18:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:29.978 15:18:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:30.237 15:18:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:30.237 15:18:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:30.237 15:18:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:30.237 15:18:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:30.237 15:18:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:30.237 15:18:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:30.237 15:18:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:30.237 15:18:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:30.237 15:18:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:30.237 15:18:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:30.237 15:18:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:30.237 "name": "raid_bdev1", 00:12:30.237 "uuid": "b40e3c58-b86f-404e-a0c1-407da15de86a", 00:12:30.237 "strip_size_kb": 0, 00:12:30.237 "state": "online", 00:12:30.237 "raid_level": "raid1", 00:12:30.237 "superblock": false, 00:12:30.237 "num_base_bdevs": 2, 00:12:30.237 "num_base_bdevs_discovered": 1, 00:12:30.237 "num_base_bdevs_operational": 1, 00:12:30.237 "base_bdevs_list": [ 00:12:30.237 { 00:12:30.237 "name": null, 00:12:30.237 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:30.237 "is_configured": false, 00:12:30.237 "data_offset": 0, 00:12:30.237 "data_size": 65536 00:12:30.237 }, 00:12:30.237 { 00:12:30.237 "name": "BaseBdev2", 00:12:30.237 "uuid": "3315fdd4-a5c2-51c0-a37b-4171bac80265", 00:12:30.237 "is_configured": true, 00:12:30.237 "data_offset": 0, 00:12:30.237 "data_size": 65536 00:12:30.237 } 00:12:30.237 ] 00:12:30.237 }' 00:12:30.237 15:18:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:30.237 15:18:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:30.237 15:18:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:30.238 15:18:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:30.238 15:18:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:30.238 15:18:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:30.238 15:18:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:30.238 [2024-11-19 15:18:20.514491] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:30.238 [2024-11-19 15:18:20.523368] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d062f0 00:12:30.238 [2024-11-19 15:18:20.525710] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:30.238 15:18:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:30.238 15:18:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:12:31.620 15:18:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:31.620 15:18:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:31.620 15:18:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:31.620 15:18:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:31.620 15:18:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:31.620 15:18:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:31.620 15:18:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:31.620 15:18:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:31.620 15:18:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:31.620 15:18:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:31.620 15:18:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:31.620 "name": "raid_bdev1", 00:12:31.620 "uuid": "b40e3c58-b86f-404e-a0c1-407da15de86a", 00:12:31.620 "strip_size_kb": 0, 00:12:31.620 "state": "online", 00:12:31.620 "raid_level": "raid1", 00:12:31.620 "superblock": false, 00:12:31.620 "num_base_bdevs": 2, 00:12:31.620 "num_base_bdevs_discovered": 2, 00:12:31.620 "num_base_bdevs_operational": 2, 00:12:31.620 "process": { 00:12:31.620 "type": "rebuild", 00:12:31.620 "target": "spare", 00:12:31.620 "progress": { 00:12:31.620 "blocks": 20480, 00:12:31.620 "percent": 31 00:12:31.620 } 00:12:31.620 }, 00:12:31.620 "base_bdevs_list": [ 00:12:31.620 { 00:12:31.620 "name": "spare", 00:12:31.620 "uuid": "05ba46d7-1719-55d6-8164-42e5b53afd28", 00:12:31.620 "is_configured": true, 00:12:31.620 "data_offset": 0, 00:12:31.620 "data_size": 65536 00:12:31.620 }, 00:12:31.620 { 00:12:31.620 "name": "BaseBdev2", 00:12:31.620 "uuid": "3315fdd4-a5c2-51c0-a37b-4171bac80265", 00:12:31.620 "is_configured": true, 00:12:31.620 "data_offset": 0, 00:12:31.620 "data_size": 65536 00:12:31.620 } 00:12:31.620 ] 00:12:31.620 }' 00:12:31.620 15:18:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:31.620 15:18:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:31.620 15:18:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:31.620 15:18:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:31.620 15:18:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:12:31.620 15:18:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:12:31.621 15:18:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:12:31.621 15:18:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:12:31.621 15:18:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=299 00:12:31.621 15:18:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:31.621 15:18:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:31.621 15:18:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:31.621 15:18:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:31.621 15:18:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:31.621 15:18:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:31.621 15:18:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:31.621 15:18:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:31.621 15:18:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:31.621 15:18:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:31.621 15:18:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:31.621 15:18:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:31.621 "name": "raid_bdev1", 00:12:31.621 "uuid": "b40e3c58-b86f-404e-a0c1-407da15de86a", 00:12:31.621 "strip_size_kb": 0, 00:12:31.621 "state": "online", 00:12:31.621 "raid_level": "raid1", 00:12:31.621 "superblock": false, 00:12:31.621 "num_base_bdevs": 2, 00:12:31.621 "num_base_bdevs_discovered": 2, 00:12:31.621 "num_base_bdevs_operational": 2, 00:12:31.621 "process": { 00:12:31.621 "type": "rebuild", 00:12:31.621 "target": "spare", 00:12:31.621 "progress": { 00:12:31.621 "blocks": 22528, 00:12:31.621 "percent": 34 00:12:31.621 } 00:12:31.621 }, 00:12:31.621 "base_bdevs_list": [ 00:12:31.621 { 00:12:31.621 "name": "spare", 00:12:31.621 "uuid": "05ba46d7-1719-55d6-8164-42e5b53afd28", 00:12:31.621 "is_configured": true, 00:12:31.621 "data_offset": 0, 00:12:31.621 "data_size": 65536 00:12:31.621 }, 00:12:31.621 { 00:12:31.621 "name": "BaseBdev2", 00:12:31.621 "uuid": "3315fdd4-a5c2-51c0-a37b-4171bac80265", 00:12:31.621 "is_configured": true, 00:12:31.621 "data_offset": 0, 00:12:31.621 "data_size": 65536 00:12:31.621 } 00:12:31.621 ] 00:12:31.621 }' 00:12:31.621 15:18:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:31.621 15:18:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:31.621 15:18:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:31.621 15:18:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:31.621 15:18:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:32.559 15:18:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:32.559 15:18:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:32.559 15:18:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:32.559 15:18:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:32.559 15:18:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:32.559 15:18:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:32.559 15:18:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:32.559 15:18:22 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:32.559 15:18:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:32.559 15:18:22 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:32.559 15:18:22 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:32.818 15:18:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:32.818 "name": "raid_bdev1", 00:12:32.818 "uuid": "b40e3c58-b86f-404e-a0c1-407da15de86a", 00:12:32.818 "strip_size_kb": 0, 00:12:32.818 "state": "online", 00:12:32.818 "raid_level": "raid1", 00:12:32.818 "superblock": false, 00:12:32.818 "num_base_bdevs": 2, 00:12:32.818 "num_base_bdevs_discovered": 2, 00:12:32.818 "num_base_bdevs_operational": 2, 00:12:32.818 "process": { 00:12:32.818 "type": "rebuild", 00:12:32.818 "target": "spare", 00:12:32.818 "progress": { 00:12:32.818 "blocks": 47104, 00:12:32.818 "percent": 71 00:12:32.818 } 00:12:32.818 }, 00:12:32.818 "base_bdevs_list": [ 00:12:32.818 { 00:12:32.818 "name": "spare", 00:12:32.818 "uuid": "05ba46d7-1719-55d6-8164-42e5b53afd28", 00:12:32.818 "is_configured": true, 00:12:32.818 "data_offset": 0, 00:12:32.818 "data_size": 65536 00:12:32.818 }, 00:12:32.818 { 00:12:32.818 "name": "BaseBdev2", 00:12:32.818 "uuid": "3315fdd4-a5c2-51c0-a37b-4171bac80265", 00:12:32.818 "is_configured": true, 00:12:32.818 "data_offset": 0, 00:12:32.818 "data_size": 65536 00:12:32.818 } 00:12:32.818 ] 00:12:32.818 }' 00:12:32.818 15:18:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:32.818 15:18:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:32.818 15:18:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:32.818 15:18:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:32.818 15:18:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:33.757 [2024-11-19 15:18:23.750844] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:12:33.757 [2024-11-19 15:18:23.750997] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:12:33.757 [2024-11-19 15:18:23.751047] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:33.757 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:33.757 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:33.757 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:33.757 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:33.757 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:33.757 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:33.757 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:33.757 15:18:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:33.757 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:33.757 15:18:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:33.757 15:18:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:33.757 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:33.757 "name": "raid_bdev1", 00:12:33.757 "uuid": "b40e3c58-b86f-404e-a0c1-407da15de86a", 00:12:33.757 "strip_size_kb": 0, 00:12:33.757 "state": "online", 00:12:33.757 "raid_level": "raid1", 00:12:33.757 "superblock": false, 00:12:33.757 "num_base_bdevs": 2, 00:12:33.757 "num_base_bdevs_discovered": 2, 00:12:33.757 "num_base_bdevs_operational": 2, 00:12:33.757 "base_bdevs_list": [ 00:12:33.757 { 00:12:33.757 "name": "spare", 00:12:33.757 "uuid": "05ba46d7-1719-55d6-8164-42e5b53afd28", 00:12:33.757 "is_configured": true, 00:12:33.757 "data_offset": 0, 00:12:33.757 "data_size": 65536 00:12:33.757 }, 00:12:33.757 { 00:12:33.757 "name": "BaseBdev2", 00:12:33.757 "uuid": "3315fdd4-a5c2-51c0-a37b-4171bac80265", 00:12:33.757 "is_configured": true, 00:12:33.757 "data_offset": 0, 00:12:33.757 "data_size": 65536 00:12:33.757 } 00:12:33.757 ] 00:12:33.757 }' 00:12:33.757 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:34.018 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:12:34.018 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:34.018 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:12:34.018 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:12:34.018 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:34.018 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:34.018 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:34.018 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:34.018 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:34.018 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:34.018 15:18:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:34.018 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:34.018 15:18:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:34.018 15:18:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:34.018 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:34.018 "name": "raid_bdev1", 00:12:34.018 "uuid": "b40e3c58-b86f-404e-a0c1-407da15de86a", 00:12:34.018 "strip_size_kb": 0, 00:12:34.018 "state": "online", 00:12:34.018 "raid_level": "raid1", 00:12:34.018 "superblock": false, 00:12:34.018 "num_base_bdevs": 2, 00:12:34.018 "num_base_bdevs_discovered": 2, 00:12:34.018 "num_base_bdevs_operational": 2, 00:12:34.018 "base_bdevs_list": [ 00:12:34.018 { 00:12:34.018 "name": "spare", 00:12:34.018 "uuid": "05ba46d7-1719-55d6-8164-42e5b53afd28", 00:12:34.018 "is_configured": true, 00:12:34.018 "data_offset": 0, 00:12:34.018 "data_size": 65536 00:12:34.018 }, 00:12:34.018 { 00:12:34.018 "name": "BaseBdev2", 00:12:34.018 "uuid": "3315fdd4-a5c2-51c0-a37b-4171bac80265", 00:12:34.018 "is_configured": true, 00:12:34.018 "data_offset": 0, 00:12:34.018 "data_size": 65536 00:12:34.018 } 00:12:34.018 ] 00:12:34.018 }' 00:12:34.018 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:34.018 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:34.018 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:34.018 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:34.018 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:34.018 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:34.018 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:34.018 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:34.018 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:34.018 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:34.018 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:34.018 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:34.018 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:34.018 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:34.018 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:34.018 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:34.018 15:18:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:34.018 15:18:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:34.018 15:18:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:34.278 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:34.278 "name": "raid_bdev1", 00:12:34.278 "uuid": "b40e3c58-b86f-404e-a0c1-407da15de86a", 00:12:34.278 "strip_size_kb": 0, 00:12:34.278 "state": "online", 00:12:34.278 "raid_level": "raid1", 00:12:34.278 "superblock": false, 00:12:34.278 "num_base_bdevs": 2, 00:12:34.278 "num_base_bdevs_discovered": 2, 00:12:34.278 "num_base_bdevs_operational": 2, 00:12:34.278 "base_bdevs_list": [ 00:12:34.278 { 00:12:34.278 "name": "spare", 00:12:34.278 "uuid": "05ba46d7-1719-55d6-8164-42e5b53afd28", 00:12:34.278 "is_configured": true, 00:12:34.278 "data_offset": 0, 00:12:34.278 "data_size": 65536 00:12:34.278 }, 00:12:34.278 { 00:12:34.278 "name": "BaseBdev2", 00:12:34.278 "uuid": "3315fdd4-a5c2-51c0-a37b-4171bac80265", 00:12:34.278 "is_configured": true, 00:12:34.278 "data_offset": 0, 00:12:34.278 "data_size": 65536 00:12:34.278 } 00:12:34.278 ] 00:12:34.278 }' 00:12:34.278 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:34.278 15:18:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:34.538 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:12:34.538 15:18:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:34.538 15:18:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:34.538 [2024-11-19 15:18:24.774062] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:34.538 [2024-11-19 15:18:24.774190] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:34.538 [2024-11-19 15:18:24.774341] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:34.538 [2024-11-19 15:18:24.774439] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:34.538 [2024-11-19 15:18:24.774548] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:12:34.538 15:18:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:34.538 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:34.538 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:12:34.538 15:18:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:34.538 15:18:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:34.538 15:18:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:34.538 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:12:34.538 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:12:34.538 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:12:34.538 15:18:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:12:34.538 15:18:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:34.538 15:18:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:12:34.538 15:18:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:34.538 15:18:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:12:34.538 15:18:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:34.538 15:18:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:12:34.538 15:18:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:34.538 15:18:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:34.538 15:18:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:12:34.798 /dev/nbd0 00:12:34.798 15:18:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:34.798 15:18:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:34.798 15:18:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:12:34.798 15:18:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:12:34.798 15:18:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:34.798 15:18:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:34.798 15:18:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:12:34.798 15:18:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@877 -- # break 00:12:34.798 15:18:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:34.798 15:18:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:34.798 15:18:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:34.798 1+0 records in 00:12:34.798 1+0 records out 00:12:34.798 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000505246 s, 8.1 MB/s 00:12:34.798 15:18:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:34.798 15:18:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:12:34.798 15:18:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:34.798 15:18:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:34.798 15:18:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:12:34.799 15:18:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:34.799 15:18:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:34.799 15:18:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:12:35.058 /dev/nbd1 00:12:35.058 15:18:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:12:35.058 15:18:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:12:35.058 15:18:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:12:35.058 15:18:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:12:35.058 15:18:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:35.058 15:18:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:35.058 15:18:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:12:35.058 15:18:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@877 -- # break 00:12:35.058 15:18:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:35.058 15:18:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:35.058 15:18:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:35.058 1+0 records in 00:12:35.058 1+0 records out 00:12:35.058 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000344726 s, 11.9 MB/s 00:12:35.058 15:18:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:35.058 15:18:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:12:35.058 15:18:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:35.058 15:18:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:35.058 15:18:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:12:35.058 15:18:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:35.058 15:18:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:35.058 15:18:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:12:35.318 15:18:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:12:35.318 15:18:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:35.318 15:18:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:12:35.318 15:18:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:35.318 15:18:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:12:35.318 15:18:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:35.318 15:18:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:35.577 15:18:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:35.577 15:18:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:35.577 15:18:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:35.577 15:18:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:35.577 15:18:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:35.577 15:18:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:35.577 15:18:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:12:35.577 15:18:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:12:35.577 15:18:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:35.577 15:18:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:12:35.577 15:18:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:12:35.577 15:18:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:12:35.577 15:18:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:12:35.577 15:18:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:35.577 15:18:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:35.577 15:18:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:12:35.577 15:18:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:12:35.577 15:18:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:12:35.577 15:18:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:12:35.577 15:18:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 86047 00:12:35.577 15:18:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@954 -- # '[' -z 86047 ']' 00:12:35.577 15:18:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@958 -- # kill -0 86047 00:12:35.577 15:18:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@959 -- # uname 00:12:35.577 15:18:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:12:35.577 15:18:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 86047 00:12:35.836 15:18:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:12:35.836 15:18:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:12:35.836 killing process with pid 86047 00:12:35.836 15:18:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 86047' 00:12:35.836 15:18:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@973 -- # kill 86047 00:12:35.836 Received shutdown signal, test time was about 60.000000 seconds 00:12:35.836 00:12:35.836 Latency(us) 00:12:35.836 [2024-11-19T15:18:26.175Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:35.836 [2024-11-19T15:18:26.175Z] =================================================================================================================== 00:12:35.836 [2024-11-19T15:18:26.175Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:12:35.836 [2024-11-19 15:18:25.924510] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:35.836 15:18:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@978 -- # wait 86047 00:12:35.836 [2024-11-19 15:18:25.983151] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:36.095 15:18:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:12:36.095 00:12:36.095 real 0m14.450s 00:12:36.095 user 0m16.128s 00:12:36.095 sys 0m3.058s 00:12:36.095 15:18:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:12:36.095 15:18:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:36.095 ************************************ 00:12:36.095 END TEST raid_rebuild_test 00:12:36.095 ************************************ 00:12:36.095 15:18:26 bdev_raid -- bdev/bdev_raid.sh@979 -- # run_test raid_rebuild_test_sb raid_rebuild_test raid1 2 true false true 00:12:36.095 15:18:26 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:12:36.095 15:18:26 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:12:36.095 15:18:26 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:36.095 ************************************ 00:12:36.095 START TEST raid_rebuild_test_sb 00:12:36.095 ************************************ 00:12:36.095 15:18:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 true false true 00:12:36.095 15:18:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:12:36.095 15:18:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:12:36.095 15:18:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:12:36.095 15:18:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:12:36.095 15:18:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:12:36.095 15:18:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:12:36.095 15:18:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:36.095 15:18:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:12:36.095 15:18:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:36.095 15:18:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:36.095 15:18:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:12:36.095 15:18:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:36.096 15:18:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:36.096 15:18:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:12:36.096 15:18:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:12:36.096 15:18:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:12:36.096 15:18:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:12:36.096 15:18:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:12:36.096 15:18:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:12:36.096 15:18:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:12:36.096 15:18:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:12:36.096 15:18:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:12:36.096 15:18:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:12:36.096 15:18:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:12:36.096 15:18:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=86460 00:12:36.096 15:18:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:12:36.096 15:18:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 86460 00:12:36.096 15:18:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@835 -- # '[' -z 86460 ']' 00:12:36.096 15:18:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:36.096 15:18:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:12:36.096 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:36.096 15:18:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:36.096 15:18:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:12:36.096 15:18:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:36.355 I/O size of 3145728 is greater than zero copy threshold (65536). 00:12:36.355 Zero copy mechanism will not be used. 00:12:36.355 [2024-11-19 15:18:26.471637] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:12:36.355 [2024-11-19 15:18:26.471794] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid86460 ] 00:12:36.355 [2024-11-19 15:18:26.623888] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:36.355 [2024-11-19 15:18:26.665160] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:12:36.615 [2024-11-19 15:18:26.741539] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:36.615 [2024-11-19 15:18:26.741584] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:37.185 15:18:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:12:37.185 15:18:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@868 -- # return 0 00:12:37.185 15:18:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:37.185 15:18:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:12:37.185 15:18:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:37.185 15:18:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:37.185 BaseBdev1_malloc 00:12:37.185 15:18:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:37.185 15:18:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:37.185 15:18:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:37.185 15:18:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:37.185 [2024-11-19 15:18:27.320123] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:37.185 [2024-11-19 15:18:27.320197] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:37.185 [2024-11-19 15:18:27.320230] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:12:37.185 [2024-11-19 15:18:27.320244] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:37.185 [2024-11-19 15:18:27.322713] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:37.185 [2024-11-19 15:18:27.322753] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:37.185 BaseBdev1 00:12:37.185 15:18:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:37.185 15:18:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:37.185 15:18:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:12:37.185 15:18:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:37.185 15:18:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:37.185 BaseBdev2_malloc 00:12:37.185 15:18:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:37.185 15:18:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:12:37.185 15:18:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:37.185 15:18:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:37.185 [2024-11-19 15:18:27.354797] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:12:37.185 [2024-11-19 15:18:27.354855] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:37.185 [2024-11-19 15:18:27.354880] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:12:37.185 [2024-11-19 15:18:27.354890] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:37.185 [2024-11-19 15:18:27.357338] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:37.185 [2024-11-19 15:18:27.357378] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:12:37.185 BaseBdev2 00:12:37.185 15:18:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:37.185 15:18:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:12:37.185 15:18:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:37.185 15:18:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:37.185 spare_malloc 00:12:37.185 15:18:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:37.185 15:18:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:12:37.185 15:18:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:37.185 15:18:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:37.185 spare_delay 00:12:37.185 15:18:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:37.185 15:18:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:37.185 15:18:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:37.185 15:18:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:37.185 [2024-11-19 15:18:27.401762] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:37.185 [2024-11-19 15:18:27.401830] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:37.185 [2024-11-19 15:18:27.401856] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:12:37.185 [2024-11-19 15:18:27.401864] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:37.185 [2024-11-19 15:18:27.404388] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:37.185 [2024-11-19 15:18:27.404422] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:37.185 spare 00:12:37.185 15:18:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:37.185 15:18:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:12:37.185 15:18:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:37.185 15:18:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:37.185 [2024-11-19 15:18:27.413797] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:37.186 [2024-11-19 15:18:27.415951] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:37.186 [2024-11-19 15:18:27.416128] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:12:37.186 [2024-11-19 15:18:27.416142] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:37.186 [2024-11-19 15:18:27.416450] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:12:37.186 [2024-11-19 15:18:27.416627] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:12:37.186 [2024-11-19 15:18:27.416648] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:12:37.186 [2024-11-19 15:18:27.416780] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:37.186 15:18:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:37.186 15:18:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:37.186 15:18:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:37.186 15:18:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:37.186 15:18:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:37.186 15:18:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:37.186 15:18:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:37.186 15:18:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:37.186 15:18:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:37.186 15:18:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:37.186 15:18:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:37.186 15:18:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:37.186 15:18:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:37.186 15:18:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:37.186 15:18:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:37.186 15:18:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:37.186 15:18:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:37.186 "name": "raid_bdev1", 00:12:37.186 "uuid": "9464bb42-c18d-4096-8151-59c635d945f6", 00:12:37.186 "strip_size_kb": 0, 00:12:37.186 "state": "online", 00:12:37.186 "raid_level": "raid1", 00:12:37.186 "superblock": true, 00:12:37.186 "num_base_bdevs": 2, 00:12:37.186 "num_base_bdevs_discovered": 2, 00:12:37.186 "num_base_bdevs_operational": 2, 00:12:37.186 "base_bdevs_list": [ 00:12:37.186 { 00:12:37.186 "name": "BaseBdev1", 00:12:37.186 "uuid": "7fe5abd1-f47c-5265-b63a-aeda3e6aa194", 00:12:37.186 "is_configured": true, 00:12:37.186 "data_offset": 2048, 00:12:37.186 "data_size": 63488 00:12:37.186 }, 00:12:37.186 { 00:12:37.186 "name": "BaseBdev2", 00:12:37.186 "uuid": "79c57729-2a13-5733-bcba-643b9581ddd0", 00:12:37.186 "is_configured": true, 00:12:37.186 "data_offset": 2048, 00:12:37.186 "data_size": 63488 00:12:37.186 } 00:12:37.186 ] 00:12:37.186 }' 00:12:37.186 15:18:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:37.186 15:18:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:37.754 15:18:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:37.754 15:18:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:12:37.754 15:18:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:37.754 15:18:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:37.754 [2024-11-19 15:18:27.849377] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:37.754 15:18:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:37.754 15:18:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:12:37.754 15:18:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:37.754 15:18:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:12:37.754 15:18:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:37.754 15:18:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:37.754 15:18:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:37.754 15:18:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:12:37.754 15:18:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:12:37.754 15:18:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:12:37.754 15:18:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:12:37.754 15:18:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:12:37.754 15:18:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:37.755 15:18:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:12:37.755 15:18:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:37.755 15:18:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:12:37.755 15:18:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:37.755 15:18:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:12:37.755 15:18:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:37.755 15:18:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:37.755 15:18:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:12:38.022 [2024-11-19 15:18:28.092769] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:12:38.022 /dev/nbd0 00:12:38.022 15:18:28 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:38.022 15:18:28 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:38.022 15:18:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:12:38.022 15:18:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:12:38.022 15:18:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:38.022 15:18:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:38.022 15:18:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:12:38.022 15:18:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:12:38.022 15:18:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:38.022 15:18:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:38.022 15:18:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:38.022 1+0 records in 00:12:38.022 1+0 records out 00:12:38.022 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000363668 s, 11.3 MB/s 00:12:38.022 15:18:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:38.022 15:18:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:12:38.022 15:18:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:38.023 15:18:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:38.023 15:18:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:12:38.023 15:18:28 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:38.023 15:18:28 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:38.023 15:18:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:12:38.023 15:18:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:12:38.023 15:18:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=63488 oflag=direct 00:12:42.234 63488+0 records in 00:12:42.234 63488+0 records out 00:12:42.234 32505856 bytes (33 MB, 31 MiB) copied, 4.24277 s, 7.7 MB/s 00:12:42.234 15:18:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:12:42.234 15:18:32 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:42.234 15:18:32 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:12:42.234 15:18:32 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:42.234 15:18:32 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:12:42.234 15:18:32 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:42.234 15:18:32 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:42.494 [2024-11-19 15:18:32.622561] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:42.494 15:18:32 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:42.494 15:18:32 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:42.494 15:18:32 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:42.494 15:18:32 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:42.494 15:18:32 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:42.494 15:18:32 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:42.494 15:18:32 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:12:42.494 15:18:32 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:12:42.494 15:18:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:12:42.494 15:18:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:42.494 15:18:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:42.494 [2024-11-19 15:18:32.646614] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:12:42.494 15:18:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:42.494 15:18:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:42.494 15:18:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:42.494 15:18:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:42.494 15:18:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:42.494 15:18:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:42.494 15:18:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:42.494 15:18:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:42.494 15:18:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:42.494 15:18:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:42.494 15:18:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:42.494 15:18:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:42.494 15:18:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:42.494 15:18:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:42.494 15:18:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:42.494 15:18:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:42.494 15:18:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:42.494 "name": "raid_bdev1", 00:12:42.494 "uuid": "9464bb42-c18d-4096-8151-59c635d945f6", 00:12:42.494 "strip_size_kb": 0, 00:12:42.494 "state": "online", 00:12:42.494 "raid_level": "raid1", 00:12:42.494 "superblock": true, 00:12:42.494 "num_base_bdevs": 2, 00:12:42.494 "num_base_bdevs_discovered": 1, 00:12:42.494 "num_base_bdevs_operational": 1, 00:12:42.494 "base_bdevs_list": [ 00:12:42.494 { 00:12:42.494 "name": null, 00:12:42.494 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:42.494 "is_configured": false, 00:12:42.494 "data_offset": 0, 00:12:42.494 "data_size": 63488 00:12:42.494 }, 00:12:42.494 { 00:12:42.494 "name": "BaseBdev2", 00:12:42.494 "uuid": "79c57729-2a13-5733-bcba-643b9581ddd0", 00:12:42.494 "is_configured": true, 00:12:42.494 "data_offset": 2048, 00:12:42.494 "data_size": 63488 00:12:42.494 } 00:12:42.494 ] 00:12:42.494 }' 00:12:42.494 15:18:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:42.494 15:18:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:43.064 15:18:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:43.064 15:18:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:43.064 15:18:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:43.064 [2024-11-19 15:18:33.097818] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:43.064 [2024-11-19 15:18:33.113024] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000c3e280 00:12:43.064 15:18:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:43.064 15:18:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:12:43.064 [2024-11-19 15:18:33.119062] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:44.004 15:18:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:44.004 15:18:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:44.004 15:18:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:44.004 15:18:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:44.004 15:18:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:44.004 15:18:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:44.004 15:18:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:44.004 15:18:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:44.004 15:18:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:44.004 15:18:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:44.004 15:18:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:44.004 "name": "raid_bdev1", 00:12:44.004 "uuid": "9464bb42-c18d-4096-8151-59c635d945f6", 00:12:44.004 "strip_size_kb": 0, 00:12:44.004 "state": "online", 00:12:44.004 "raid_level": "raid1", 00:12:44.004 "superblock": true, 00:12:44.004 "num_base_bdevs": 2, 00:12:44.004 "num_base_bdevs_discovered": 2, 00:12:44.004 "num_base_bdevs_operational": 2, 00:12:44.004 "process": { 00:12:44.004 "type": "rebuild", 00:12:44.004 "target": "spare", 00:12:44.004 "progress": { 00:12:44.004 "blocks": 20480, 00:12:44.004 "percent": 32 00:12:44.004 } 00:12:44.004 }, 00:12:44.004 "base_bdevs_list": [ 00:12:44.004 { 00:12:44.004 "name": "spare", 00:12:44.004 "uuid": "cd8b05c7-77e6-5f60-a366-c6d069c58041", 00:12:44.004 "is_configured": true, 00:12:44.004 "data_offset": 2048, 00:12:44.004 "data_size": 63488 00:12:44.004 }, 00:12:44.004 { 00:12:44.004 "name": "BaseBdev2", 00:12:44.004 "uuid": "79c57729-2a13-5733-bcba-643b9581ddd0", 00:12:44.004 "is_configured": true, 00:12:44.004 "data_offset": 2048, 00:12:44.004 "data_size": 63488 00:12:44.004 } 00:12:44.004 ] 00:12:44.004 }' 00:12:44.004 15:18:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:44.004 15:18:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:44.004 15:18:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:44.004 15:18:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:44.004 15:18:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:44.004 15:18:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:44.004 15:18:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:44.004 [2024-11-19 15:18:34.245819] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:44.004 [2024-11-19 15:18:34.324144] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:44.004 [2024-11-19 15:18:34.324222] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:44.004 [2024-11-19 15:18:34.324243] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:44.004 [2024-11-19 15:18:34.324251] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:44.004 15:18:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:44.004 15:18:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:44.004 15:18:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:44.004 15:18:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:44.004 15:18:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:44.004 15:18:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:44.004 15:18:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:44.004 15:18:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:44.004 15:18:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:44.004 15:18:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:44.004 15:18:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:44.264 15:18:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:44.264 15:18:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:44.264 15:18:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:44.264 15:18:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:44.264 15:18:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:44.264 15:18:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:44.264 "name": "raid_bdev1", 00:12:44.264 "uuid": "9464bb42-c18d-4096-8151-59c635d945f6", 00:12:44.264 "strip_size_kb": 0, 00:12:44.265 "state": "online", 00:12:44.265 "raid_level": "raid1", 00:12:44.265 "superblock": true, 00:12:44.265 "num_base_bdevs": 2, 00:12:44.265 "num_base_bdevs_discovered": 1, 00:12:44.265 "num_base_bdevs_operational": 1, 00:12:44.265 "base_bdevs_list": [ 00:12:44.265 { 00:12:44.265 "name": null, 00:12:44.265 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:44.265 "is_configured": false, 00:12:44.265 "data_offset": 0, 00:12:44.265 "data_size": 63488 00:12:44.265 }, 00:12:44.265 { 00:12:44.265 "name": "BaseBdev2", 00:12:44.265 "uuid": "79c57729-2a13-5733-bcba-643b9581ddd0", 00:12:44.265 "is_configured": true, 00:12:44.265 "data_offset": 2048, 00:12:44.265 "data_size": 63488 00:12:44.265 } 00:12:44.265 ] 00:12:44.265 }' 00:12:44.265 15:18:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:44.265 15:18:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:44.524 15:18:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:44.525 15:18:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:44.525 15:18:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:44.525 15:18:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:44.525 15:18:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:44.525 15:18:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:44.525 15:18:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:44.525 15:18:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:44.525 15:18:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:44.525 15:18:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:44.525 15:18:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:44.525 "name": "raid_bdev1", 00:12:44.525 "uuid": "9464bb42-c18d-4096-8151-59c635d945f6", 00:12:44.525 "strip_size_kb": 0, 00:12:44.525 "state": "online", 00:12:44.525 "raid_level": "raid1", 00:12:44.525 "superblock": true, 00:12:44.525 "num_base_bdevs": 2, 00:12:44.525 "num_base_bdevs_discovered": 1, 00:12:44.525 "num_base_bdevs_operational": 1, 00:12:44.525 "base_bdevs_list": [ 00:12:44.525 { 00:12:44.525 "name": null, 00:12:44.525 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:44.525 "is_configured": false, 00:12:44.525 "data_offset": 0, 00:12:44.525 "data_size": 63488 00:12:44.525 }, 00:12:44.525 { 00:12:44.525 "name": "BaseBdev2", 00:12:44.525 "uuid": "79c57729-2a13-5733-bcba-643b9581ddd0", 00:12:44.525 "is_configured": true, 00:12:44.525 "data_offset": 2048, 00:12:44.525 "data_size": 63488 00:12:44.525 } 00:12:44.525 ] 00:12:44.525 }' 00:12:44.525 15:18:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:44.525 15:18:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:44.525 15:18:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:44.525 15:18:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:44.525 15:18:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:44.525 15:18:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:44.525 15:18:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:44.784 [2024-11-19 15:18:34.868091] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:44.784 [2024-11-19 15:18:34.872984] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000c3e350 00:12:44.784 15:18:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:44.784 15:18:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:12:44.784 [2024-11-19 15:18:34.874853] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:45.767 15:18:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:45.767 15:18:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:45.767 15:18:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:45.767 15:18:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:45.767 15:18:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:45.767 15:18:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:45.767 15:18:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:45.767 15:18:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:45.767 15:18:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:45.767 15:18:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:45.767 15:18:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:45.767 "name": "raid_bdev1", 00:12:45.767 "uuid": "9464bb42-c18d-4096-8151-59c635d945f6", 00:12:45.767 "strip_size_kb": 0, 00:12:45.767 "state": "online", 00:12:45.767 "raid_level": "raid1", 00:12:45.767 "superblock": true, 00:12:45.767 "num_base_bdevs": 2, 00:12:45.767 "num_base_bdevs_discovered": 2, 00:12:45.767 "num_base_bdevs_operational": 2, 00:12:45.767 "process": { 00:12:45.767 "type": "rebuild", 00:12:45.767 "target": "spare", 00:12:45.767 "progress": { 00:12:45.767 "blocks": 20480, 00:12:45.767 "percent": 32 00:12:45.767 } 00:12:45.767 }, 00:12:45.767 "base_bdevs_list": [ 00:12:45.767 { 00:12:45.767 "name": "spare", 00:12:45.767 "uuid": "cd8b05c7-77e6-5f60-a366-c6d069c58041", 00:12:45.767 "is_configured": true, 00:12:45.767 "data_offset": 2048, 00:12:45.767 "data_size": 63488 00:12:45.767 }, 00:12:45.767 { 00:12:45.767 "name": "BaseBdev2", 00:12:45.767 "uuid": "79c57729-2a13-5733-bcba-643b9581ddd0", 00:12:45.767 "is_configured": true, 00:12:45.767 "data_offset": 2048, 00:12:45.767 "data_size": 63488 00:12:45.767 } 00:12:45.767 ] 00:12:45.767 }' 00:12:45.767 15:18:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:45.767 15:18:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:45.767 15:18:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:45.767 15:18:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:45.767 15:18:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:12:45.767 15:18:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:12:45.767 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:12:45.767 15:18:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:12:45.767 15:18:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:12:45.767 15:18:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:12:45.767 15:18:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=314 00:12:45.767 15:18:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:45.767 15:18:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:45.767 15:18:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:45.767 15:18:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:45.767 15:18:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:45.767 15:18:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:45.767 15:18:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:45.767 15:18:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:45.767 15:18:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:45.767 15:18:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:45.767 15:18:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:45.767 15:18:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:45.767 "name": "raid_bdev1", 00:12:45.767 "uuid": "9464bb42-c18d-4096-8151-59c635d945f6", 00:12:45.767 "strip_size_kb": 0, 00:12:45.767 "state": "online", 00:12:45.767 "raid_level": "raid1", 00:12:45.768 "superblock": true, 00:12:45.768 "num_base_bdevs": 2, 00:12:45.768 "num_base_bdevs_discovered": 2, 00:12:45.768 "num_base_bdevs_operational": 2, 00:12:45.768 "process": { 00:12:45.768 "type": "rebuild", 00:12:45.768 "target": "spare", 00:12:45.768 "progress": { 00:12:45.768 "blocks": 22528, 00:12:45.768 "percent": 35 00:12:45.768 } 00:12:45.768 }, 00:12:45.768 "base_bdevs_list": [ 00:12:45.768 { 00:12:45.768 "name": "spare", 00:12:45.768 "uuid": "cd8b05c7-77e6-5f60-a366-c6d069c58041", 00:12:45.768 "is_configured": true, 00:12:45.768 "data_offset": 2048, 00:12:45.768 "data_size": 63488 00:12:45.768 }, 00:12:45.768 { 00:12:45.768 "name": "BaseBdev2", 00:12:45.768 "uuid": "79c57729-2a13-5733-bcba-643b9581ddd0", 00:12:45.768 "is_configured": true, 00:12:45.768 "data_offset": 2048, 00:12:45.768 "data_size": 63488 00:12:45.768 } 00:12:45.768 ] 00:12:45.768 }' 00:12:45.768 15:18:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:45.768 15:18:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:45.768 15:18:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:46.027 15:18:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:46.027 15:18:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:46.966 15:18:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:46.966 15:18:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:46.966 15:18:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:46.966 15:18:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:46.966 15:18:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:46.966 15:18:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:46.966 15:18:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:46.966 15:18:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:46.966 15:18:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:46.966 15:18:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:46.966 15:18:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:46.966 15:18:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:46.966 "name": "raid_bdev1", 00:12:46.966 "uuid": "9464bb42-c18d-4096-8151-59c635d945f6", 00:12:46.966 "strip_size_kb": 0, 00:12:46.966 "state": "online", 00:12:46.966 "raid_level": "raid1", 00:12:46.966 "superblock": true, 00:12:46.966 "num_base_bdevs": 2, 00:12:46.966 "num_base_bdevs_discovered": 2, 00:12:46.966 "num_base_bdevs_operational": 2, 00:12:46.966 "process": { 00:12:46.966 "type": "rebuild", 00:12:46.966 "target": "spare", 00:12:46.966 "progress": { 00:12:46.966 "blocks": 45056, 00:12:46.966 "percent": 70 00:12:46.966 } 00:12:46.966 }, 00:12:46.966 "base_bdevs_list": [ 00:12:46.966 { 00:12:46.966 "name": "spare", 00:12:46.966 "uuid": "cd8b05c7-77e6-5f60-a366-c6d069c58041", 00:12:46.966 "is_configured": true, 00:12:46.966 "data_offset": 2048, 00:12:46.966 "data_size": 63488 00:12:46.966 }, 00:12:46.966 { 00:12:46.966 "name": "BaseBdev2", 00:12:46.966 "uuid": "79c57729-2a13-5733-bcba-643b9581ddd0", 00:12:46.966 "is_configured": true, 00:12:46.966 "data_offset": 2048, 00:12:46.966 "data_size": 63488 00:12:46.966 } 00:12:46.966 ] 00:12:46.966 }' 00:12:46.966 15:18:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:46.966 15:18:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:46.966 15:18:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:46.966 15:18:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:46.966 15:18:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:47.905 [2024-11-19 15:18:37.985039] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:12:47.905 [2024-11-19 15:18:37.985113] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:12:47.905 [2024-11-19 15:18:37.985215] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:48.164 15:18:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:48.164 15:18:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:48.164 15:18:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:48.164 15:18:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:48.164 15:18:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:48.164 15:18:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:48.164 15:18:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:48.164 15:18:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:48.164 15:18:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:48.164 15:18:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:48.164 15:18:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:48.164 15:18:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:48.164 "name": "raid_bdev1", 00:12:48.164 "uuid": "9464bb42-c18d-4096-8151-59c635d945f6", 00:12:48.164 "strip_size_kb": 0, 00:12:48.164 "state": "online", 00:12:48.164 "raid_level": "raid1", 00:12:48.164 "superblock": true, 00:12:48.164 "num_base_bdevs": 2, 00:12:48.164 "num_base_bdevs_discovered": 2, 00:12:48.164 "num_base_bdevs_operational": 2, 00:12:48.164 "base_bdevs_list": [ 00:12:48.164 { 00:12:48.164 "name": "spare", 00:12:48.164 "uuid": "cd8b05c7-77e6-5f60-a366-c6d069c58041", 00:12:48.164 "is_configured": true, 00:12:48.164 "data_offset": 2048, 00:12:48.164 "data_size": 63488 00:12:48.164 }, 00:12:48.164 { 00:12:48.164 "name": "BaseBdev2", 00:12:48.164 "uuid": "79c57729-2a13-5733-bcba-643b9581ddd0", 00:12:48.164 "is_configured": true, 00:12:48.164 "data_offset": 2048, 00:12:48.164 "data_size": 63488 00:12:48.164 } 00:12:48.164 ] 00:12:48.164 }' 00:12:48.164 15:18:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:48.164 15:18:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:12:48.164 15:18:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:48.164 15:18:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:12:48.164 15:18:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:12:48.164 15:18:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:48.164 15:18:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:48.164 15:18:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:48.165 15:18:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:48.165 15:18:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:48.165 15:18:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:48.165 15:18:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:48.165 15:18:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:48.165 15:18:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:48.165 15:18:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:48.165 15:18:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:48.165 "name": "raid_bdev1", 00:12:48.165 "uuid": "9464bb42-c18d-4096-8151-59c635d945f6", 00:12:48.165 "strip_size_kb": 0, 00:12:48.165 "state": "online", 00:12:48.165 "raid_level": "raid1", 00:12:48.165 "superblock": true, 00:12:48.165 "num_base_bdevs": 2, 00:12:48.165 "num_base_bdevs_discovered": 2, 00:12:48.165 "num_base_bdevs_operational": 2, 00:12:48.165 "base_bdevs_list": [ 00:12:48.165 { 00:12:48.165 "name": "spare", 00:12:48.165 "uuid": "cd8b05c7-77e6-5f60-a366-c6d069c58041", 00:12:48.165 "is_configured": true, 00:12:48.165 "data_offset": 2048, 00:12:48.165 "data_size": 63488 00:12:48.165 }, 00:12:48.165 { 00:12:48.165 "name": "BaseBdev2", 00:12:48.165 "uuid": "79c57729-2a13-5733-bcba-643b9581ddd0", 00:12:48.165 "is_configured": true, 00:12:48.165 "data_offset": 2048, 00:12:48.165 "data_size": 63488 00:12:48.165 } 00:12:48.165 ] 00:12:48.165 }' 00:12:48.165 15:18:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:48.424 15:18:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:48.424 15:18:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:48.424 15:18:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:48.424 15:18:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:48.424 15:18:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:48.424 15:18:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:48.424 15:18:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:48.424 15:18:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:48.424 15:18:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:48.424 15:18:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:48.424 15:18:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:48.424 15:18:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:48.424 15:18:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:48.424 15:18:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:48.424 15:18:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:48.424 15:18:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:48.424 15:18:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:48.424 15:18:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:48.424 15:18:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:48.424 "name": "raid_bdev1", 00:12:48.424 "uuid": "9464bb42-c18d-4096-8151-59c635d945f6", 00:12:48.424 "strip_size_kb": 0, 00:12:48.424 "state": "online", 00:12:48.424 "raid_level": "raid1", 00:12:48.425 "superblock": true, 00:12:48.425 "num_base_bdevs": 2, 00:12:48.425 "num_base_bdevs_discovered": 2, 00:12:48.425 "num_base_bdevs_operational": 2, 00:12:48.425 "base_bdevs_list": [ 00:12:48.425 { 00:12:48.425 "name": "spare", 00:12:48.425 "uuid": "cd8b05c7-77e6-5f60-a366-c6d069c58041", 00:12:48.425 "is_configured": true, 00:12:48.425 "data_offset": 2048, 00:12:48.425 "data_size": 63488 00:12:48.425 }, 00:12:48.425 { 00:12:48.425 "name": "BaseBdev2", 00:12:48.425 "uuid": "79c57729-2a13-5733-bcba-643b9581ddd0", 00:12:48.425 "is_configured": true, 00:12:48.425 "data_offset": 2048, 00:12:48.425 "data_size": 63488 00:12:48.425 } 00:12:48.425 ] 00:12:48.425 }' 00:12:48.425 15:18:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:48.425 15:18:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:48.685 15:18:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:12:48.685 15:18:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:48.685 15:18:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:48.685 [2024-11-19 15:18:38.963918] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:48.685 [2024-11-19 15:18:38.964010] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:48.685 [2024-11-19 15:18:38.964129] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:48.685 [2024-11-19 15:18:38.964220] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:48.685 [2024-11-19 15:18:38.964266] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:12:48.685 15:18:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:48.685 15:18:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:48.685 15:18:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:12:48.685 15:18:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:48.685 15:18:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:48.685 15:18:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:48.685 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:12:48.685 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:12:48.685 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:12:48.685 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:12:48.685 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:48.685 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:12:48.685 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:48.685 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:12:48.685 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:48.685 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:12:48.685 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:48.685 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:48.685 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:12:48.945 /dev/nbd0 00:12:48.945 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:48.945 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:48.945 15:18:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:12:48.945 15:18:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:12:48.945 15:18:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:48.945 15:18:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:48.945 15:18:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:12:48.945 15:18:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:12:48.945 15:18:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:48.945 15:18:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:48.945 15:18:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:48.945 1+0 records in 00:12:48.945 1+0 records out 00:12:48.945 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000542457 s, 7.6 MB/s 00:12:48.945 15:18:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:48.945 15:18:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:12:48.945 15:18:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:48.945 15:18:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:48.945 15:18:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:12:48.945 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:48.945 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:48.945 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:12:49.204 /dev/nbd1 00:12:49.204 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:12:49.204 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:12:49.204 15:18:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:12:49.204 15:18:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:12:49.204 15:18:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:49.204 15:18:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:49.204 15:18:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:12:49.204 15:18:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:12:49.204 15:18:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:49.205 15:18:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:49.205 15:18:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:49.205 1+0 records in 00:12:49.205 1+0 records out 00:12:49.205 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000390428 s, 10.5 MB/s 00:12:49.205 15:18:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:49.205 15:18:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:12:49.205 15:18:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:49.205 15:18:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:49.205 15:18:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:12:49.205 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:49.205 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:49.205 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:12:49.465 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:12:49.465 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:49.465 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:12:49.465 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:49.465 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:12:49.465 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:49.465 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:49.465 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:49.465 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:49.465 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:49.465 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:49.465 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:49.465 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:49.465 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:12:49.465 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:12:49.465 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:49.465 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:12:49.725 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:12:49.725 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:12:49.725 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:12:49.725 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:49.725 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:49.725 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:12:49.725 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:12:49.725 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:12:49.725 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:12:49.725 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:12:49.725 15:18:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:49.725 15:18:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:49.725 15:18:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:49.725 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:49.725 15:18:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:49.725 15:18:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:49.725 [2024-11-19 15:18:39.980955] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:49.725 [2024-11-19 15:18:39.981021] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:49.725 [2024-11-19 15:18:39.981042] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:12:49.725 [2024-11-19 15:18:39.981055] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:49.725 [2024-11-19 15:18:39.983044] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:49.725 [2024-11-19 15:18:39.983081] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:49.725 [2024-11-19 15:18:39.983163] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:12:49.725 [2024-11-19 15:18:39.983215] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:49.725 [2024-11-19 15:18:39.983347] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:49.725 spare 00:12:49.725 15:18:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:49.725 15:18:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:12:49.725 15:18:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:49.725 15:18:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:49.987 [2024-11-19 15:18:40.083246] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:12:49.987 [2024-11-19 15:18:40.083308] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:49.987 [2024-11-19 15:18:40.083599] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000cae960 00:12:49.987 [2024-11-19 15:18:40.083760] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:12:49.987 [2024-11-19 15:18:40.083774] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:12:49.987 [2024-11-19 15:18:40.083906] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:49.987 15:18:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:49.987 15:18:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:49.987 15:18:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:49.987 15:18:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:49.987 15:18:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:49.987 15:18:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:49.987 15:18:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:49.987 15:18:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:49.987 15:18:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:49.987 15:18:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:49.987 15:18:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:49.987 15:18:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:49.987 15:18:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:49.987 15:18:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:49.987 15:18:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:49.987 15:18:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:49.987 15:18:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:49.987 "name": "raid_bdev1", 00:12:49.987 "uuid": "9464bb42-c18d-4096-8151-59c635d945f6", 00:12:49.987 "strip_size_kb": 0, 00:12:49.987 "state": "online", 00:12:49.987 "raid_level": "raid1", 00:12:49.987 "superblock": true, 00:12:49.987 "num_base_bdevs": 2, 00:12:49.987 "num_base_bdevs_discovered": 2, 00:12:49.987 "num_base_bdevs_operational": 2, 00:12:49.987 "base_bdevs_list": [ 00:12:49.987 { 00:12:49.987 "name": "spare", 00:12:49.987 "uuid": "cd8b05c7-77e6-5f60-a366-c6d069c58041", 00:12:49.987 "is_configured": true, 00:12:49.987 "data_offset": 2048, 00:12:49.987 "data_size": 63488 00:12:49.987 }, 00:12:49.987 { 00:12:49.987 "name": "BaseBdev2", 00:12:49.987 "uuid": "79c57729-2a13-5733-bcba-643b9581ddd0", 00:12:49.987 "is_configured": true, 00:12:49.987 "data_offset": 2048, 00:12:49.987 "data_size": 63488 00:12:49.987 } 00:12:49.987 ] 00:12:49.987 }' 00:12:49.987 15:18:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:49.987 15:18:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:50.247 15:18:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:50.247 15:18:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:50.247 15:18:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:50.247 15:18:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:50.247 15:18:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:50.247 15:18:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:50.247 15:18:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:50.247 15:18:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:50.247 15:18:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:50.507 15:18:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:50.507 15:18:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:50.507 "name": "raid_bdev1", 00:12:50.507 "uuid": "9464bb42-c18d-4096-8151-59c635d945f6", 00:12:50.507 "strip_size_kb": 0, 00:12:50.507 "state": "online", 00:12:50.507 "raid_level": "raid1", 00:12:50.507 "superblock": true, 00:12:50.507 "num_base_bdevs": 2, 00:12:50.507 "num_base_bdevs_discovered": 2, 00:12:50.507 "num_base_bdevs_operational": 2, 00:12:50.507 "base_bdevs_list": [ 00:12:50.507 { 00:12:50.507 "name": "spare", 00:12:50.507 "uuid": "cd8b05c7-77e6-5f60-a366-c6d069c58041", 00:12:50.507 "is_configured": true, 00:12:50.507 "data_offset": 2048, 00:12:50.507 "data_size": 63488 00:12:50.507 }, 00:12:50.507 { 00:12:50.507 "name": "BaseBdev2", 00:12:50.507 "uuid": "79c57729-2a13-5733-bcba-643b9581ddd0", 00:12:50.507 "is_configured": true, 00:12:50.507 "data_offset": 2048, 00:12:50.507 "data_size": 63488 00:12:50.507 } 00:12:50.507 ] 00:12:50.507 }' 00:12:50.507 15:18:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:50.507 15:18:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:50.507 15:18:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:50.507 15:18:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:50.507 15:18:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:50.507 15:18:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:12:50.507 15:18:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:50.507 15:18:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:50.507 15:18:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:50.507 15:18:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:12:50.507 15:18:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:50.507 15:18:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:50.507 15:18:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:50.507 [2024-11-19 15:18:40.783710] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:50.507 15:18:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:50.507 15:18:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:50.507 15:18:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:50.507 15:18:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:50.507 15:18:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:50.507 15:18:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:50.507 15:18:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:50.507 15:18:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:50.507 15:18:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:50.507 15:18:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:50.507 15:18:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:50.507 15:18:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:50.507 15:18:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:50.507 15:18:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:50.507 15:18:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:50.507 15:18:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:50.507 15:18:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:50.507 "name": "raid_bdev1", 00:12:50.507 "uuid": "9464bb42-c18d-4096-8151-59c635d945f6", 00:12:50.507 "strip_size_kb": 0, 00:12:50.507 "state": "online", 00:12:50.507 "raid_level": "raid1", 00:12:50.507 "superblock": true, 00:12:50.507 "num_base_bdevs": 2, 00:12:50.507 "num_base_bdevs_discovered": 1, 00:12:50.507 "num_base_bdevs_operational": 1, 00:12:50.507 "base_bdevs_list": [ 00:12:50.507 { 00:12:50.507 "name": null, 00:12:50.507 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:50.507 "is_configured": false, 00:12:50.507 "data_offset": 0, 00:12:50.507 "data_size": 63488 00:12:50.507 }, 00:12:50.507 { 00:12:50.507 "name": "BaseBdev2", 00:12:50.507 "uuid": "79c57729-2a13-5733-bcba-643b9581ddd0", 00:12:50.507 "is_configured": true, 00:12:50.507 "data_offset": 2048, 00:12:50.507 "data_size": 63488 00:12:50.507 } 00:12:50.507 ] 00:12:50.507 }' 00:12:50.507 15:18:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:50.507 15:18:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:51.077 15:18:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:51.077 15:18:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:51.077 15:18:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:51.077 [2024-11-19 15:18:41.179038] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:51.077 [2024-11-19 15:18:41.179215] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:12:51.077 [2024-11-19 15:18:41.179230] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:12:51.077 [2024-11-19 15:18:41.179284] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:51.077 [2024-11-19 15:18:41.184247] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000caea30 00:12:51.077 15:18:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:51.077 15:18:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:12:51.077 [2024-11-19 15:18:41.186075] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:52.017 15:18:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:52.017 15:18:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:52.017 15:18:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:52.017 15:18:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:52.017 15:18:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:52.017 15:18:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:52.017 15:18:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:52.017 15:18:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:52.017 15:18:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:52.017 15:18:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:52.017 15:18:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:52.017 "name": "raid_bdev1", 00:12:52.017 "uuid": "9464bb42-c18d-4096-8151-59c635d945f6", 00:12:52.017 "strip_size_kb": 0, 00:12:52.017 "state": "online", 00:12:52.017 "raid_level": "raid1", 00:12:52.017 "superblock": true, 00:12:52.017 "num_base_bdevs": 2, 00:12:52.017 "num_base_bdevs_discovered": 2, 00:12:52.017 "num_base_bdevs_operational": 2, 00:12:52.017 "process": { 00:12:52.017 "type": "rebuild", 00:12:52.017 "target": "spare", 00:12:52.017 "progress": { 00:12:52.017 "blocks": 20480, 00:12:52.017 "percent": 32 00:12:52.017 } 00:12:52.017 }, 00:12:52.017 "base_bdevs_list": [ 00:12:52.017 { 00:12:52.017 "name": "spare", 00:12:52.017 "uuid": "cd8b05c7-77e6-5f60-a366-c6d069c58041", 00:12:52.017 "is_configured": true, 00:12:52.017 "data_offset": 2048, 00:12:52.017 "data_size": 63488 00:12:52.017 }, 00:12:52.017 { 00:12:52.017 "name": "BaseBdev2", 00:12:52.017 "uuid": "79c57729-2a13-5733-bcba-643b9581ddd0", 00:12:52.017 "is_configured": true, 00:12:52.017 "data_offset": 2048, 00:12:52.017 "data_size": 63488 00:12:52.017 } 00:12:52.017 ] 00:12:52.017 }' 00:12:52.017 15:18:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:52.017 15:18:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:52.017 15:18:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:52.017 15:18:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:52.017 15:18:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:12:52.017 15:18:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:52.017 15:18:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:52.017 [2024-11-19 15:18:42.326149] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:52.278 [2024-11-19 15:18:42.389997] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:52.278 [2024-11-19 15:18:42.390056] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:52.278 [2024-11-19 15:18:42.390071] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:52.278 [2024-11-19 15:18:42.390079] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:52.278 15:18:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:52.278 15:18:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:52.278 15:18:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:52.278 15:18:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:52.278 15:18:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:52.278 15:18:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:52.278 15:18:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:52.278 15:18:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:52.278 15:18:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:52.278 15:18:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:52.278 15:18:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:52.278 15:18:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:52.278 15:18:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:52.278 15:18:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:52.278 15:18:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:52.278 15:18:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:52.278 15:18:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:52.278 "name": "raid_bdev1", 00:12:52.278 "uuid": "9464bb42-c18d-4096-8151-59c635d945f6", 00:12:52.278 "strip_size_kb": 0, 00:12:52.278 "state": "online", 00:12:52.278 "raid_level": "raid1", 00:12:52.278 "superblock": true, 00:12:52.278 "num_base_bdevs": 2, 00:12:52.278 "num_base_bdevs_discovered": 1, 00:12:52.278 "num_base_bdevs_operational": 1, 00:12:52.278 "base_bdevs_list": [ 00:12:52.278 { 00:12:52.278 "name": null, 00:12:52.278 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:52.278 "is_configured": false, 00:12:52.278 "data_offset": 0, 00:12:52.278 "data_size": 63488 00:12:52.278 }, 00:12:52.278 { 00:12:52.278 "name": "BaseBdev2", 00:12:52.278 "uuid": "79c57729-2a13-5733-bcba-643b9581ddd0", 00:12:52.278 "is_configured": true, 00:12:52.278 "data_offset": 2048, 00:12:52.278 "data_size": 63488 00:12:52.278 } 00:12:52.278 ] 00:12:52.278 }' 00:12:52.278 15:18:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:52.278 15:18:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:52.538 15:18:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:52.538 15:18:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:52.538 15:18:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:52.538 [2024-11-19 15:18:42.854011] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:52.538 [2024-11-19 15:18:42.854131] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:52.538 [2024-11-19 15:18:42.854173] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:12:52.538 [2024-11-19 15:18:42.854205] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:52.538 [2024-11-19 15:18:42.854668] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:52.538 [2024-11-19 15:18:42.854686] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:52.538 [2024-11-19 15:18:42.854768] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:12:52.538 [2024-11-19 15:18:42.854780] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:12:52.538 [2024-11-19 15:18:42.854796] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:12:52.538 [2024-11-19 15:18:42.854824] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:52.538 [2024-11-19 15:18:42.859562] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000caeb00 00:12:52.538 spare 00:12:52.538 15:18:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:52.538 15:18:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:12:52.538 [2024-11-19 15:18:42.861453] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:53.922 15:18:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:53.922 15:18:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:53.922 15:18:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:53.922 15:18:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:53.922 15:18:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:53.922 15:18:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:53.922 15:18:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:53.922 15:18:43 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:53.922 15:18:43 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:53.922 15:18:43 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:53.922 15:18:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:53.922 "name": "raid_bdev1", 00:12:53.922 "uuid": "9464bb42-c18d-4096-8151-59c635d945f6", 00:12:53.922 "strip_size_kb": 0, 00:12:53.922 "state": "online", 00:12:53.922 "raid_level": "raid1", 00:12:53.922 "superblock": true, 00:12:53.922 "num_base_bdevs": 2, 00:12:53.922 "num_base_bdevs_discovered": 2, 00:12:53.922 "num_base_bdevs_operational": 2, 00:12:53.922 "process": { 00:12:53.922 "type": "rebuild", 00:12:53.922 "target": "spare", 00:12:53.922 "progress": { 00:12:53.922 "blocks": 20480, 00:12:53.922 "percent": 32 00:12:53.922 } 00:12:53.922 }, 00:12:53.922 "base_bdevs_list": [ 00:12:53.922 { 00:12:53.922 "name": "spare", 00:12:53.922 "uuid": "cd8b05c7-77e6-5f60-a366-c6d069c58041", 00:12:53.922 "is_configured": true, 00:12:53.922 "data_offset": 2048, 00:12:53.922 "data_size": 63488 00:12:53.922 }, 00:12:53.922 { 00:12:53.922 "name": "BaseBdev2", 00:12:53.922 "uuid": "79c57729-2a13-5733-bcba-643b9581ddd0", 00:12:53.922 "is_configured": true, 00:12:53.922 "data_offset": 2048, 00:12:53.922 "data_size": 63488 00:12:53.922 } 00:12:53.922 ] 00:12:53.922 }' 00:12:53.922 15:18:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:53.922 15:18:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:53.922 15:18:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:53.922 15:18:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:53.922 15:18:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:12:53.922 15:18:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:53.922 15:18:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:53.922 [2024-11-19 15:18:44.021568] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:53.922 [2024-11-19 15:18:44.065411] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:53.922 [2024-11-19 15:18:44.065469] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:53.922 [2024-11-19 15:18:44.065484] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:53.922 [2024-11-19 15:18:44.065493] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:53.922 15:18:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:53.922 15:18:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:53.922 15:18:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:53.922 15:18:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:53.922 15:18:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:53.922 15:18:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:53.922 15:18:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:53.922 15:18:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:53.922 15:18:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:53.922 15:18:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:53.922 15:18:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:53.922 15:18:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:53.922 15:18:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:53.922 15:18:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:53.922 15:18:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:53.922 15:18:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:53.922 15:18:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:53.922 "name": "raid_bdev1", 00:12:53.922 "uuid": "9464bb42-c18d-4096-8151-59c635d945f6", 00:12:53.922 "strip_size_kb": 0, 00:12:53.922 "state": "online", 00:12:53.922 "raid_level": "raid1", 00:12:53.922 "superblock": true, 00:12:53.922 "num_base_bdevs": 2, 00:12:53.922 "num_base_bdevs_discovered": 1, 00:12:53.922 "num_base_bdevs_operational": 1, 00:12:53.922 "base_bdevs_list": [ 00:12:53.922 { 00:12:53.922 "name": null, 00:12:53.922 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:53.922 "is_configured": false, 00:12:53.922 "data_offset": 0, 00:12:53.922 "data_size": 63488 00:12:53.922 }, 00:12:53.922 { 00:12:53.922 "name": "BaseBdev2", 00:12:53.922 "uuid": "79c57729-2a13-5733-bcba-643b9581ddd0", 00:12:53.922 "is_configured": true, 00:12:53.922 "data_offset": 2048, 00:12:53.922 "data_size": 63488 00:12:53.922 } 00:12:53.922 ] 00:12:53.922 }' 00:12:53.922 15:18:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:53.922 15:18:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:54.492 15:18:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:54.492 15:18:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:54.492 15:18:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:54.492 15:18:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:54.492 15:18:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:54.492 15:18:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:54.492 15:18:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:54.492 15:18:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:54.492 15:18:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:54.492 15:18:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:54.492 15:18:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:54.492 "name": "raid_bdev1", 00:12:54.492 "uuid": "9464bb42-c18d-4096-8151-59c635d945f6", 00:12:54.492 "strip_size_kb": 0, 00:12:54.492 "state": "online", 00:12:54.492 "raid_level": "raid1", 00:12:54.492 "superblock": true, 00:12:54.492 "num_base_bdevs": 2, 00:12:54.492 "num_base_bdevs_discovered": 1, 00:12:54.492 "num_base_bdevs_operational": 1, 00:12:54.492 "base_bdevs_list": [ 00:12:54.492 { 00:12:54.492 "name": null, 00:12:54.492 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:54.492 "is_configured": false, 00:12:54.492 "data_offset": 0, 00:12:54.492 "data_size": 63488 00:12:54.492 }, 00:12:54.492 { 00:12:54.492 "name": "BaseBdev2", 00:12:54.492 "uuid": "79c57729-2a13-5733-bcba-643b9581ddd0", 00:12:54.492 "is_configured": true, 00:12:54.492 "data_offset": 2048, 00:12:54.492 "data_size": 63488 00:12:54.492 } 00:12:54.492 ] 00:12:54.492 }' 00:12:54.492 15:18:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:54.492 15:18:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:54.492 15:18:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:54.492 15:18:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:54.492 15:18:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:12:54.492 15:18:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:54.492 15:18:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:54.492 15:18:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:54.492 15:18:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:54.492 15:18:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:54.492 15:18:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:54.492 [2024-11-19 15:18:44.700905] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:54.492 [2024-11-19 15:18:44.700974] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:54.492 [2024-11-19 15:18:44.700996] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:12:54.492 [2024-11-19 15:18:44.701006] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:54.492 [2024-11-19 15:18:44.701431] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:54.492 [2024-11-19 15:18:44.701456] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:54.492 [2024-11-19 15:18:44.701528] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:12:54.492 [2024-11-19 15:18:44.701545] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:12:54.492 [2024-11-19 15:18:44.701553] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:12:54.492 [2024-11-19 15:18:44.701564] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:12:54.492 BaseBdev1 00:12:54.492 15:18:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:54.492 15:18:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:12:55.432 15:18:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:55.432 15:18:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:55.432 15:18:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:55.432 15:18:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:55.432 15:18:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:55.432 15:18:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:55.432 15:18:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:55.432 15:18:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:55.432 15:18:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:55.432 15:18:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:55.432 15:18:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:55.432 15:18:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:55.432 15:18:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:55.432 15:18:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:55.432 15:18:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:55.432 15:18:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:55.432 "name": "raid_bdev1", 00:12:55.432 "uuid": "9464bb42-c18d-4096-8151-59c635d945f6", 00:12:55.432 "strip_size_kb": 0, 00:12:55.432 "state": "online", 00:12:55.432 "raid_level": "raid1", 00:12:55.432 "superblock": true, 00:12:55.432 "num_base_bdevs": 2, 00:12:55.432 "num_base_bdevs_discovered": 1, 00:12:55.432 "num_base_bdevs_operational": 1, 00:12:55.432 "base_bdevs_list": [ 00:12:55.432 { 00:12:55.432 "name": null, 00:12:55.432 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:55.432 "is_configured": false, 00:12:55.432 "data_offset": 0, 00:12:55.432 "data_size": 63488 00:12:55.432 }, 00:12:55.432 { 00:12:55.432 "name": "BaseBdev2", 00:12:55.432 "uuid": "79c57729-2a13-5733-bcba-643b9581ddd0", 00:12:55.432 "is_configured": true, 00:12:55.432 "data_offset": 2048, 00:12:55.432 "data_size": 63488 00:12:55.432 } 00:12:55.432 ] 00:12:55.432 }' 00:12:55.432 15:18:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:55.432 15:18:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:56.003 15:18:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:56.003 15:18:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:56.003 15:18:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:56.003 15:18:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:56.003 15:18:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:56.003 15:18:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:56.003 15:18:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:56.003 15:18:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:56.003 15:18:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:56.003 15:18:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:56.003 15:18:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:56.003 "name": "raid_bdev1", 00:12:56.003 "uuid": "9464bb42-c18d-4096-8151-59c635d945f6", 00:12:56.003 "strip_size_kb": 0, 00:12:56.003 "state": "online", 00:12:56.003 "raid_level": "raid1", 00:12:56.003 "superblock": true, 00:12:56.003 "num_base_bdevs": 2, 00:12:56.003 "num_base_bdevs_discovered": 1, 00:12:56.003 "num_base_bdevs_operational": 1, 00:12:56.003 "base_bdevs_list": [ 00:12:56.003 { 00:12:56.003 "name": null, 00:12:56.003 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:56.003 "is_configured": false, 00:12:56.003 "data_offset": 0, 00:12:56.003 "data_size": 63488 00:12:56.003 }, 00:12:56.003 { 00:12:56.003 "name": "BaseBdev2", 00:12:56.003 "uuid": "79c57729-2a13-5733-bcba-643b9581ddd0", 00:12:56.003 "is_configured": true, 00:12:56.003 "data_offset": 2048, 00:12:56.003 "data_size": 63488 00:12:56.003 } 00:12:56.003 ] 00:12:56.003 }' 00:12:56.003 15:18:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:56.003 15:18:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:56.003 15:18:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:56.003 15:18:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:56.003 15:18:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:56.003 15:18:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@652 -- # local es=0 00:12:56.003 15:18:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:56.003 15:18:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:12:56.003 15:18:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:12:56.003 15:18:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:12:56.003 15:18:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:12:56.003 15:18:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:56.003 15:18:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:56.003 15:18:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:56.003 [2024-11-19 15:18:46.294715] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:56.003 [2024-11-19 15:18:46.294936] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:12:56.003 [2024-11-19 15:18:46.294953] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:12:56.003 request: 00:12:56.003 { 00:12:56.003 "base_bdev": "BaseBdev1", 00:12:56.003 "raid_bdev": "raid_bdev1", 00:12:56.003 "method": "bdev_raid_add_base_bdev", 00:12:56.003 "req_id": 1 00:12:56.003 } 00:12:56.003 Got JSON-RPC error response 00:12:56.003 response: 00:12:56.003 { 00:12:56.003 "code": -22, 00:12:56.003 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:12:56.003 } 00:12:56.003 15:18:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:12:56.003 15:18:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@655 -- # es=1 00:12:56.003 15:18:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:12:56.003 15:18:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:12:56.003 15:18:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:12:56.003 15:18:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:12:57.383 15:18:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:57.383 15:18:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:57.383 15:18:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:57.383 15:18:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:57.383 15:18:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:57.383 15:18:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:57.383 15:18:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:57.383 15:18:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:57.383 15:18:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:57.383 15:18:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:57.383 15:18:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:57.383 15:18:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:57.383 15:18:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:57.383 15:18:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:57.383 15:18:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:57.383 15:18:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:57.383 "name": "raid_bdev1", 00:12:57.383 "uuid": "9464bb42-c18d-4096-8151-59c635d945f6", 00:12:57.383 "strip_size_kb": 0, 00:12:57.383 "state": "online", 00:12:57.383 "raid_level": "raid1", 00:12:57.383 "superblock": true, 00:12:57.383 "num_base_bdevs": 2, 00:12:57.383 "num_base_bdevs_discovered": 1, 00:12:57.383 "num_base_bdevs_operational": 1, 00:12:57.383 "base_bdevs_list": [ 00:12:57.383 { 00:12:57.383 "name": null, 00:12:57.383 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:57.383 "is_configured": false, 00:12:57.383 "data_offset": 0, 00:12:57.383 "data_size": 63488 00:12:57.383 }, 00:12:57.383 { 00:12:57.383 "name": "BaseBdev2", 00:12:57.383 "uuid": "79c57729-2a13-5733-bcba-643b9581ddd0", 00:12:57.383 "is_configured": true, 00:12:57.383 "data_offset": 2048, 00:12:57.383 "data_size": 63488 00:12:57.383 } 00:12:57.383 ] 00:12:57.383 }' 00:12:57.383 15:18:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:57.383 15:18:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:57.643 15:18:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:57.643 15:18:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:57.643 15:18:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:57.643 15:18:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:57.643 15:18:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:57.643 15:18:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:57.643 15:18:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:57.643 15:18:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:57.643 15:18:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:57.643 15:18:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:57.643 15:18:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:57.643 "name": "raid_bdev1", 00:12:57.643 "uuid": "9464bb42-c18d-4096-8151-59c635d945f6", 00:12:57.643 "strip_size_kb": 0, 00:12:57.643 "state": "online", 00:12:57.643 "raid_level": "raid1", 00:12:57.643 "superblock": true, 00:12:57.643 "num_base_bdevs": 2, 00:12:57.643 "num_base_bdevs_discovered": 1, 00:12:57.643 "num_base_bdevs_operational": 1, 00:12:57.643 "base_bdevs_list": [ 00:12:57.643 { 00:12:57.643 "name": null, 00:12:57.643 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:57.643 "is_configured": false, 00:12:57.643 "data_offset": 0, 00:12:57.643 "data_size": 63488 00:12:57.643 }, 00:12:57.643 { 00:12:57.643 "name": "BaseBdev2", 00:12:57.643 "uuid": "79c57729-2a13-5733-bcba-643b9581ddd0", 00:12:57.643 "is_configured": true, 00:12:57.643 "data_offset": 2048, 00:12:57.643 "data_size": 63488 00:12:57.643 } 00:12:57.643 ] 00:12:57.643 }' 00:12:57.643 15:18:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:57.643 15:18:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:57.643 15:18:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:57.643 15:18:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:57.643 15:18:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 86460 00:12:57.643 15:18:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@954 -- # '[' -z 86460 ']' 00:12:57.643 15:18:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@958 -- # kill -0 86460 00:12:57.643 15:18:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@959 -- # uname 00:12:57.643 15:18:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:12:57.643 15:18:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 86460 00:12:57.643 killing process with pid 86460 00:12:57.643 Received shutdown signal, test time was about 60.000000 seconds 00:12:57.643 00:12:57.643 Latency(us) 00:12:57.643 [2024-11-19T15:18:47.982Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:57.643 [2024-11-19T15:18:47.982Z] =================================================================================================================== 00:12:57.643 [2024-11-19T15:18:47.982Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:12:57.643 15:18:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:12:57.643 15:18:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:12:57.643 15:18:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 86460' 00:12:57.643 15:18:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@973 -- # kill 86460 00:12:57.643 [2024-11-19 15:18:47.902324] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:57.643 [2024-11-19 15:18:47.902452] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:57.643 15:18:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@978 -- # wait 86460 00:12:57.643 [2024-11-19 15:18:47.902505] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:57.643 [2024-11-19 15:18:47.902514] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:12:57.643 [2024-11-19 15:18:47.933761] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:57.903 15:18:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:12:57.903 00:12:57.903 real 0m21.757s 00:12:57.903 user 0m26.588s 00:12:57.903 sys 0m3.839s 00:12:57.903 15:18:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:12:57.903 15:18:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:57.903 ************************************ 00:12:57.903 END TEST raid_rebuild_test_sb 00:12:57.903 ************************************ 00:12:57.903 15:18:48 bdev_raid -- bdev/bdev_raid.sh@980 -- # run_test raid_rebuild_test_io raid_rebuild_test raid1 2 false true true 00:12:57.903 15:18:48 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:12:57.903 15:18:48 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:12:57.903 15:18:48 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:57.903 ************************************ 00:12:57.903 START TEST raid_rebuild_test_io 00:12:57.903 ************************************ 00:12:57.903 15:18:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 false true true 00:12:57.903 15:18:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:12:57.903 15:18:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:12:57.903 15:18:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:12:57.903 15:18:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:12:57.903 15:18:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:12:57.903 15:18:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:12:57.903 15:18:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:57.903 15:18:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:12:57.903 15:18:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:57.903 15:18:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:57.903 15:18:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:12:57.903 15:18:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:57.903 15:18:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:57.903 15:18:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:12:57.903 15:18:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:12:57.903 15:18:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:12:57.903 15:18:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:12:57.903 15:18:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:12:57.903 15:18:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:12:57.903 15:18:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:12:57.903 15:18:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:12:57.904 15:18:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:12:57.904 15:18:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:12:57.904 15:18:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@597 -- # raid_pid=87176 00:12:57.904 15:18:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:12:57.904 15:18:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 87176 00:12:57.904 15:18:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@835 -- # '[' -z 87176 ']' 00:12:57.904 15:18:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:57.904 15:18:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@840 -- # local max_retries=100 00:12:57.904 15:18:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:57.904 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:57.904 15:18:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@844 -- # xtrace_disable 00:12:57.904 15:18:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:58.164 [2024-11-19 15:18:48.302513] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:12:58.164 [2024-11-19 15:18:48.302730] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid87176 ] 00:12:58.164 I/O size of 3145728 is greater than zero copy threshold (65536). 00:12:58.164 Zero copy mechanism will not be used. 00:12:58.164 [2024-11-19 15:18:48.458082] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:58.164 [2024-11-19 15:18:48.482822] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:12:58.424 [2024-11-19 15:18:48.524971] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:58.424 [2024-11-19 15:18:48.525091] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:58.997 15:18:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:12:58.997 15:18:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@868 -- # return 0 00:12:58.997 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:58.997 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:12:58.997 15:18:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:58.997 15:18:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:58.997 BaseBdev1_malloc 00:12:58.997 15:18:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:58.997 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:58.997 15:18:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:58.997 15:18:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:58.997 [2024-11-19 15:18:49.126748] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:58.997 [2024-11-19 15:18:49.126856] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:58.997 [2024-11-19 15:18:49.126908] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:12:58.997 [2024-11-19 15:18:49.126939] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:58.997 [2024-11-19 15:18:49.129089] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:58.997 [2024-11-19 15:18:49.129178] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:58.997 BaseBdev1 00:12:58.997 15:18:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:58.997 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:58.997 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:12:58.997 15:18:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:58.997 15:18:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:58.997 BaseBdev2_malloc 00:12:58.998 15:18:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:58.998 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:12:58.998 15:18:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:58.998 15:18:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:58.998 [2024-11-19 15:18:49.155071] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:12:58.998 [2024-11-19 15:18:49.155172] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:58.998 [2024-11-19 15:18:49.155207] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:12:58.998 [2024-11-19 15:18:49.155233] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:58.998 [2024-11-19 15:18:49.157278] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:58.998 [2024-11-19 15:18:49.157368] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:12:58.998 BaseBdev2 00:12:58.998 15:18:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:58.998 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:12:58.998 15:18:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:58.998 15:18:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:58.998 spare_malloc 00:12:58.998 15:18:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:58.998 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:12:58.998 15:18:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:58.998 15:18:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:58.998 spare_delay 00:12:58.998 15:18:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:58.998 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:58.998 15:18:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:58.998 15:18:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:58.998 [2024-11-19 15:18:49.195430] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:58.998 [2024-11-19 15:18:49.195479] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:58.998 [2024-11-19 15:18:49.195515] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:12:58.998 [2024-11-19 15:18:49.195523] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:58.998 [2024-11-19 15:18:49.197576] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:58.998 [2024-11-19 15:18:49.197612] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:58.999 spare 00:12:58.999 15:18:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:58.999 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:12:58.999 15:18:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:58.999 15:18:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:58.999 [2024-11-19 15:18:49.207447] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:58.999 [2024-11-19 15:18:49.209249] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:58.999 [2024-11-19 15:18:49.209333] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:12:58.999 [2024-11-19 15:18:49.209343] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:12:58.999 [2024-11-19 15:18:49.209589] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:12:58.999 [2024-11-19 15:18:49.209722] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:12:58.999 [2024-11-19 15:18:49.209735] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:12:58.999 [2024-11-19 15:18:49.209858] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:58.999 15:18:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:58.999 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:58.999 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:58.999 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:58.999 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:58.999 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:58.999 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:58.999 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:58.999 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:58.999 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:58.999 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:58.999 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:58.999 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:58.999 15:18:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:58.999 15:18:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:58.999 15:18:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:58.999 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:58.999 "name": "raid_bdev1", 00:12:58.999 "uuid": "bad7764f-d0c1-4239-9c77-e8bbb5344d80", 00:12:58.999 "strip_size_kb": 0, 00:12:58.999 "state": "online", 00:12:58.999 "raid_level": "raid1", 00:12:58.999 "superblock": false, 00:12:58.999 "num_base_bdevs": 2, 00:12:58.999 "num_base_bdevs_discovered": 2, 00:12:58.999 "num_base_bdevs_operational": 2, 00:12:58.999 "base_bdevs_list": [ 00:12:58.999 { 00:12:58.999 "name": "BaseBdev1", 00:12:58.999 "uuid": "f5e8d26f-2515-5488-957b-04137f331980", 00:12:58.999 "is_configured": true, 00:12:58.999 "data_offset": 0, 00:12:58.999 "data_size": 65536 00:12:58.999 }, 00:12:59.000 { 00:12:59.000 "name": "BaseBdev2", 00:12:59.000 "uuid": "17977eca-470a-5f45-aa38-6f2649a1941e", 00:12:59.000 "is_configured": true, 00:12:59.000 "data_offset": 0, 00:12:59.000 "data_size": 65536 00:12:59.000 } 00:12:59.000 ] 00:12:59.000 }' 00:12:59.000 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:59.000 15:18:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:59.570 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:59.570 15:18:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:59.570 15:18:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:59.570 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:12:59.570 [2024-11-19 15:18:49.619000] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:59.570 15:18:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:59.570 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:12:59.570 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:59.570 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:12:59.570 15:18:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:59.570 15:18:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:59.570 15:18:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:59.570 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:12:59.570 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:12:59.570 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:12:59.570 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:12:59.570 15:18:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:59.570 15:18:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:59.570 [2024-11-19 15:18:49.718545] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:12:59.570 15:18:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:59.570 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:59.570 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:59.570 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:59.570 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:59.570 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:59.570 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:59.570 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:59.570 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:59.570 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:59.570 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:59.570 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:59.570 15:18:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:59.570 15:18:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:59.570 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:59.570 15:18:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:59.570 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:59.570 "name": "raid_bdev1", 00:12:59.570 "uuid": "bad7764f-d0c1-4239-9c77-e8bbb5344d80", 00:12:59.570 "strip_size_kb": 0, 00:12:59.570 "state": "online", 00:12:59.570 "raid_level": "raid1", 00:12:59.570 "superblock": false, 00:12:59.570 "num_base_bdevs": 2, 00:12:59.570 "num_base_bdevs_discovered": 1, 00:12:59.570 "num_base_bdevs_operational": 1, 00:12:59.570 "base_bdevs_list": [ 00:12:59.570 { 00:12:59.570 "name": null, 00:12:59.570 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:59.570 "is_configured": false, 00:12:59.570 "data_offset": 0, 00:12:59.570 "data_size": 65536 00:12:59.570 }, 00:12:59.571 { 00:12:59.571 "name": "BaseBdev2", 00:12:59.571 "uuid": "17977eca-470a-5f45-aa38-6f2649a1941e", 00:12:59.571 "is_configured": true, 00:12:59.571 "data_offset": 0, 00:12:59.571 "data_size": 65536 00:12:59.571 } 00:12:59.571 ] 00:12:59.571 }' 00:12:59.571 15:18:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:59.571 15:18:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:59.571 [2024-11-19 15:18:49.804380] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:12:59.571 I/O size of 3145728 is greater than zero copy threshold (65536). 00:12:59.571 Zero copy mechanism will not be used. 00:12:59.571 Running I/O for 60 seconds... 00:12:59.830 15:18:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:59.830 15:18:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:59.830 15:18:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:59.830 [2024-11-19 15:18:50.161360] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:00.089 15:18:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:00.089 15:18:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:13:00.089 [2024-11-19 15:18:50.225268] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:13:00.089 [2024-11-19 15:18:50.227259] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:00.089 [2024-11-19 15:18:50.334206] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:13:00.089 [2024-11-19 15:18:50.334562] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:13:00.348 [2024-11-19 15:18:50.553136] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:13:00.348 [2024-11-19 15:18:50.553373] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:13:00.867 189.00 IOPS, 567.00 MiB/s [2024-11-19T15:18:51.206Z] [2024-11-19 15:18:50.992951] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:13:00.867 15:18:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:00.867 15:18:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:00.867 15:18:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:00.867 15:18:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:00.867 15:18:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:01.127 15:18:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:01.127 15:18:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:01.127 15:18:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:01.127 15:18:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:01.127 15:18:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:01.127 15:18:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:01.127 "name": "raid_bdev1", 00:13:01.127 "uuid": "bad7764f-d0c1-4239-9c77-e8bbb5344d80", 00:13:01.127 "strip_size_kb": 0, 00:13:01.127 "state": "online", 00:13:01.127 "raid_level": "raid1", 00:13:01.127 "superblock": false, 00:13:01.127 "num_base_bdevs": 2, 00:13:01.127 "num_base_bdevs_discovered": 2, 00:13:01.127 "num_base_bdevs_operational": 2, 00:13:01.127 "process": { 00:13:01.127 "type": "rebuild", 00:13:01.127 "target": "spare", 00:13:01.127 "progress": { 00:13:01.127 "blocks": 12288, 00:13:01.127 "percent": 18 00:13:01.127 } 00:13:01.127 }, 00:13:01.127 "base_bdevs_list": [ 00:13:01.127 { 00:13:01.127 "name": "spare", 00:13:01.127 "uuid": "3f934860-0301-5e89-bf66-18b3cf7c62da", 00:13:01.127 "is_configured": true, 00:13:01.127 "data_offset": 0, 00:13:01.127 "data_size": 65536 00:13:01.127 }, 00:13:01.127 { 00:13:01.127 "name": "BaseBdev2", 00:13:01.127 "uuid": "17977eca-470a-5f45-aa38-6f2649a1941e", 00:13:01.127 "is_configured": true, 00:13:01.127 "data_offset": 0, 00:13:01.127 "data_size": 65536 00:13:01.127 } 00:13:01.127 ] 00:13:01.127 }' 00:13:01.127 15:18:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:01.127 15:18:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:01.127 15:18:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:01.127 [2024-11-19 15:18:51.315948] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:13:01.127 15:18:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:01.127 15:18:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:13:01.127 15:18:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:01.127 15:18:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:01.127 [2024-11-19 15:18:51.359085] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:01.127 [2024-11-19 15:18:51.450391] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:13:01.127 [2024-11-19 15:18:51.450550] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:13:01.127 [2024-11-19 15:18:51.456513] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:01.388 [2024-11-19 15:18:51.468673] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:01.388 [2024-11-19 15:18:51.468770] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:01.388 [2024-11-19 15:18:51.468786] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:01.388 [2024-11-19 15:18:51.490227] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d0000026d0 00:13:01.388 15:18:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:01.388 15:18:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:13:01.388 15:18:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:01.388 15:18:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:01.388 15:18:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:01.388 15:18:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:01.388 15:18:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:13:01.388 15:18:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:01.388 15:18:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:01.388 15:18:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:01.388 15:18:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:01.388 15:18:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:01.388 15:18:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:01.388 15:18:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:01.388 15:18:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:01.388 15:18:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:01.388 15:18:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:01.388 "name": "raid_bdev1", 00:13:01.388 "uuid": "bad7764f-d0c1-4239-9c77-e8bbb5344d80", 00:13:01.388 "strip_size_kb": 0, 00:13:01.388 "state": "online", 00:13:01.388 "raid_level": "raid1", 00:13:01.388 "superblock": false, 00:13:01.388 "num_base_bdevs": 2, 00:13:01.388 "num_base_bdevs_discovered": 1, 00:13:01.388 "num_base_bdevs_operational": 1, 00:13:01.388 "base_bdevs_list": [ 00:13:01.388 { 00:13:01.388 "name": null, 00:13:01.388 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:01.388 "is_configured": false, 00:13:01.388 "data_offset": 0, 00:13:01.388 "data_size": 65536 00:13:01.388 }, 00:13:01.388 { 00:13:01.388 "name": "BaseBdev2", 00:13:01.388 "uuid": "17977eca-470a-5f45-aa38-6f2649a1941e", 00:13:01.388 "is_configured": true, 00:13:01.388 "data_offset": 0, 00:13:01.388 "data_size": 65536 00:13:01.388 } 00:13:01.388 ] 00:13:01.388 }' 00:13:01.388 15:18:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:01.388 15:18:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:01.648 184.50 IOPS, 553.50 MiB/s [2024-11-19T15:18:51.987Z] 15:18:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:01.648 15:18:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:01.648 15:18:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:01.648 15:18:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:01.648 15:18:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:01.648 15:18:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:01.648 15:18:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:01.648 15:18:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:01.648 15:18:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:01.648 15:18:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:01.648 15:18:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:01.648 "name": "raid_bdev1", 00:13:01.648 "uuid": "bad7764f-d0c1-4239-9c77-e8bbb5344d80", 00:13:01.648 "strip_size_kb": 0, 00:13:01.648 "state": "online", 00:13:01.648 "raid_level": "raid1", 00:13:01.648 "superblock": false, 00:13:01.648 "num_base_bdevs": 2, 00:13:01.648 "num_base_bdevs_discovered": 1, 00:13:01.648 "num_base_bdevs_operational": 1, 00:13:01.648 "base_bdevs_list": [ 00:13:01.648 { 00:13:01.648 "name": null, 00:13:01.648 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:01.648 "is_configured": false, 00:13:01.648 "data_offset": 0, 00:13:01.648 "data_size": 65536 00:13:01.648 }, 00:13:01.648 { 00:13:01.648 "name": "BaseBdev2", 00:13:01.648 "uuid": "17977eca-470a-5f45-aa38-6f2649a1941e", 00:13:01.648 "is_configured": true, 00:13:01.648 "data_offset": 0, 00:13:01.648 "data_size": 65536 00:13:01.648 } 00:13:01.648 ] 00:13:01.648 }' 00:13:01.648 15:18:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:01.908 15:18:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:01.908 15:18:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:01.908 15:18:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:01.908 15:18:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:01.908 15:18:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:01.908 15:18:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:01.908 [2024-11-19 15:18:52.056051] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:01.908 15:18:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:01.908 15:18:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:13:01.908 [2024-11-19 15:18:52.077413] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:13:01.908 [2024-11-19 15:18:52.079333] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:01.908 [2024-11-19 15:18:52.207405] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:13:02.168 [2024-11-19 15:18:52.321823] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:13:02.168 [2024-11-19 15:18:52.322147] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:13:02.737 [2024-11-19 15:18:52.793797] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:13:02.997 191.67 IOPS, 575.00 MiB/s [2024-11-19T15:18:53.336Z] 15:18:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:02.997 15:18:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:02.997 15:18:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:02.997 15:18:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:02.997 15:18:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:02.997 15:18:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:02.997 15:18:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:02.997 15:18:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:02.997 15:18:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:02.997 15:18:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:02.997 15:18:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:02.997 "name": "raid_bdev1", 00:13:02.997 "uuid": "bad7764f-d0c1-4239-9c77-e8bbb5344d80", 00:13:02.997 "strip_size_kb": 0, 00:13:02.997 "state": "online", 00:13:02.997 "raid_level": "raid1", 00:13:02.997 "superblock": false, 00:13:02.997 "num_base_bdevs": 2, 00:13:02.997 "num_base_bdevs_discovered": 2, 00:13:02.997 "num_base_bdevs_operational": 2, 00:13:02.997 "process": { 00:13:02.997 "type": "rebuild", 00:13:02.997 "target": "spare", 00:13:02.997 "progress": { 00:13:02.997 "blocks": 14336, 00:13:02.997 "percent": 21 00:13:02.997 } 00:13:02.997 }, 00:13:02.997 "base_bdevs_list": [ 00:13:02.997 { 00:13:02.997 "name": "spare", 00:13:02.997 "uuid": "3f934860-0301-5e89-bf66-18b3cf7c62da", 00:13:02.997 "is_configured": true, 00:13:02.997 "data_offset": 0, 00:13:02.997 "data_size": 65536 00:13:02.997 }, 00:13:02.997 { 00:13:02.997 "name": "BaseBdev2", 00:13:02.997 "uuid": "17977eca-470a-5f45-aa38-6f2649a1941e", 00:13:02.997 "is_configured": true, 00:13:02.997 "data_offset": 0, 00:13:02.997 "data_size": 65536 00:13:02.997 } 00:13:02.997 ] 00:13:02.997 }' 00:13:02.997 15:18:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:02.997 15:18:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:02.997 15:18:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:02.997 15:18:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:02.997 15:18:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:13:02.997 15:18:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:13:02.997 15:18:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:13:02.997 15:18:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:13:02.997 15:18:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@706 -- # local timeout=331 00:13:02.997 15:18:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:02.997 15:18:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:02.997 15:18:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:02.997 15:18:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:02.997 15:18:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:02.997 15:18:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:02.997 15:18:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:02.997 15:18:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:02.997 15:18:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:02.997 15:18:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:02.997 15:18:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:02.997 15:18:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:02.997 "name": "raid_bdev1", 00:13:02.997 "uuid": "bad7764f-d0c1-4239-9c77-e8bbb5344d80", 00:13:02.997 "strip_size_kb": 0, 00:13:02.997 "state": "online", 00:13:02.997 "raid_level": "raid1", 00:13:02.997 "superblock": false, 00:13:02.997 "num_base_bdevs": 2, 00:13:02.997 "num_base_bdevs_discovered": 2, 00:13:02.997 "num_base_bdevs_operational": 2, 00:13:02.997 "process": { 00:13:02.997 "type": "rebuild", 00:13:02.997 "target": "spare", 00:13:02.997 "progress": { 00:13:02.997 "blocks": 18432, 00:13:02.997 "percent": 28 00:13:02.997 } 00:13:02.997 }, 00:13:02.997 "base_bdevs_list": [ 00:13:02.997 { 00:13:02.997 "name": "spare", 00:13:02.997 "uuid": "3f934860-0301-5e89-bf66-18b3cf7c62da", 00:13:02.997 "is_configured": true, 00:13:02.997 "data_offset": 0, 00:13:02.997 "data_size": 65536 00:13:02.997 }, 00:13:02.997 { 00:13:02.997 "name": "BaseBdev2", 00:13:02.997 "uuid": "17977eca-470a-5f45-aa38-6f2649a1941e", 00:13:02.997 "is_configured": true, 00:13:02.997 "data_offset": 0, 00:13:02.997 "data_size": 65536 00:13:02.997 } 00:13:02.997 ] 00:13:02.997 }' 00:13:02.997 15:18:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:02.997 15:18:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:02.997 15:18:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:02.997 15:18:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:02.997 15:18:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:03.257 [2024-11-19 15:18:53.415944] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:13:03.257 [2024-11-19 15:18:53.416192] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:13:03.517 [2024-11-19 15:18:53.742074] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 26624 offset_begin: 24576 offset_end: 30720 00:13:03.517 [2024-11-19 15:18:53.742561] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 26624 offset_begin: 24576 offset_end: 30720 00:13:03.776 158.00 IOPS, 474.00 MiB/s [2024-11-19T15:18:54.115Z] [2024-11-19 15:18:53.953789] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 28672 offset_begin: 24576 offset_end: 30720 00:13:03.776 [2024-11-19 15:18:53.954044] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 28672 offset_begin: 24576 offset_end: 30720 00:13:04.036 15:18:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:04.036 15:18:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:04.036 15:18:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:04.036 15:18:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:04.036 15:18:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:04.036 15:18:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:04.036 15:18:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:04.036 15:18:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:04.036 15:18:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:04.036 15:18:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:04.036 15:18:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:04.296 15:18:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:04.296 "name": "raid_bdev1", 00:13:04.296 "uuid": "bad7764f-d0c1-4239-9c77-e8bbb5344d80", 00:13:04.296 "strip_size_kb": 0, 00:13:04.296 "state": "online", 00:13:04.296 "raid_level": "raid1", 00:13:04.296 "superblock": false, 00:13:04.296 "num_base_bdevs": 2, 00:13:04.296 "num_base_bdevs_discovered": 2, 00:13:04.296 "num_base_bdevs_operational": 2, 00:13:04.296 "process": { 00:13:04.296 "type": "rebuild", 00:13:04.296 "target": "spare", 00:13:04.296 "progress": { 00:13:04.296 "blocks": 32768, 00:13:04.296 "percent": 50 00:13:04.296 } 00:13:04.296 }, 00:13:04.296 "base_bdevs_list": [ 00:13:04.296 { 00:13:04.296 "name": "spare", 00:13:04.296 "uuid": "3f934860-0301-5e89-bf66-18b3cf7c62da", 00:13:04.296 "is_configured": true, 00:13:04.296 "data_offset": 0, 00:13:04.296 "data_size": 65536 00:13:04.296 }, 00:13:04.296 { 00:13:04.296 "name": "BaseBdev2", 00:13:04.296 "uuid": "17977eca-470a-5f45-aa38-6f2649a1941e", 00:13:04.296 "is_configured": true, 00:13:04.296 "data_offset": 0, 00:13:04.296 "data_size": 65536 00:13:04.296 } 00:13:04.296 ] 00:13:04.296 }' 00:13:04.296 15:18:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:04.296 15:18:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:04.296 15:18:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:04.296 15:18:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:04.296 15:18:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:04.296 [2024-11-19 15:18:54.633565] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 38912 offset_begin: 36864 offset_end: 43008 00:13:04.556 [2024-11-19 15:18:54.744230] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 40960 offset_begin: 36864 offset_end: 43008 00:13:04.556 [2024-11-19 15:18:54.744494] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 40960 offset_begin: 36864 offset_end: 43008 00:13:04.815 132.20 IOPS, 396.60 MiB/s [2024-11-19T15:18:55.154Z] [2024-11-19 15:18:55.150161] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 47104 offset_begin: 43008 offset_end: 49152 00:13:05.384 [2024-11-19 15:18:55.468780] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 51200 offset_begin: 49152 offset_end: 55296 00:13:05.384 15:18:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:05.384 15:18:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:05.384 15:18:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:05.384 15:18:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:05.384 15:18:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:05.384 15:18:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:05.384 15:18:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:05.385 15:18:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:05.385 15:18:55 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:05.385 15:18:55 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:05.385 15:18:55 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:05.385 15:18:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:05.385 "name": "raid_bdev1", 00:13:05.385 "uuid": "bad7764f-d0c1-4239-9c77-e8bbb5344d80", 00:13:05.385 "strip_size_kb": 0, 00:13:05.385 "state": "online", 00:13:05.385 "raid_level": "raid1", 00:13:05.385 "superblock": false, 00:13:05.385 "num_base_bdevs": 2, 00:13:05.385 "num_base_bdevs_discovered": 2, 00:13:05.385 "num_base_bdevs_operational": 2, 00:13:05.385 "process": { 00:13:05.385 "type": "rebuild", 00:13:05.385 "target": "spare", 00:13:05.385 "progress": { 00:13:05.385 "blocks": 51200, 00:13:05.385 "percent": 78 00:13:05.385 } 00:13:05.385 }, 00:13:05.385 "base_bdevs_list": [ 00:13:05.385 { 00:13:05.385 "name": "spare", 00:13:05.385 "uuid": "3f934860-0301-5e89-bf66-18b3cf7c62da", 00:13:05.385 "is_configured": true, 00:13:05.385 "data_offset": 0, 00:13:05.385 "data_size": 65536 00:13:05.385 }, 00:13:05.385 { 00:13:05.385 "name": "BaseBdev2", 00:13:05.385 "uuid": "17977eca-470a-5f45-aa38-6f2649a1941e", 00:13:05.385 "is_configured": true, 00:13:05.385 "data_offset": 0, 00:13:05.385 "data_size": 65536 00:13:05.385 } 00:13:05.385 ] 00:13:05.385 }' 00:13:05.385 15:18:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:05.385 [2024-11-19 15:18:55.570584] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 53248 offset_begin: 49152 offset_end: 55296 00:13:05.385 [2024-11-19 15:18:55.570916] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 53248 offset_begin: 49152 offset_end: 55296 00:13:05.385 15:18:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:05.385 15:18:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:05.385 15:18:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:05.385 15:18:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:06.214 116.33 IOPS, 349.00 MiB/s [2024-11-19T15:18:56.553Z] [2024-11-19 15:18:56.322483] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:13:06.214 [2024-11-19 15:18:56.422300] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:13:06.214 [2024-11-19 15:18:56.423763] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:06.474 15:18:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:06.474 15:18:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:06.474 15:18:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:06.474 15:18:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:06.474 15:18:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:06.474 15:18:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:06.474 15:18:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:06.474 15:18:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:06.474 15:18:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:06.474 15:18:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:06.474 15:18:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:06.474 15:18:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:06.474 "name": "raid_bdev1", 00:13:06.474 "uuid": "bad7764f-d0c1-4239-9c77-e8bbb5344d80", 00:13:06.474 "strip_size_kb": 0, 00:13:06.474 "state": "online", 00:13:06.474 "raid_level": "raid1", 00:13:06.474 "superblock": false, 00:13:06.474 "num_base_bdevs": 2, 00:13:06.474 "num_base_bdevs_discovered": 2, 00:13:06.474 "num_base_bdevs_operational": 2, 00:13:06.474 "base_bdevs_list": [ 00:13:06.474 { 00:13:06.474 "name": "spare", 00:13:06.474 "uuid": "3f934860-0301-5e89-bf66-18b3cf7c62da", 00:13:06.474 "is_configured": true, 00:13:06.474 "data_offset": 0, 00:13:06.474 "data_size": 65536 00:13:06.474 }, 00:13:06.474 { 00:13:06.474 "name": "BaseBdev2", 00:13:06.474 "uuid": "17977eca-470a-5f45-aa38-6f2649a1941e", 00:13:06.474 "is_configured": true, 00:13:06.474 "data_offset": 0, 00:13:06.474 "data_size": 65536 00:13:06.474 } 00:13:06.474 ] 00:13:06.474 }' 00:13:06.474 15:18:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:06.474 15:18:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:13:06.474 15:18:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:06.474 15:18:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:13:06.474 15:18:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@709 -- # break 00:13:06.474 15:18:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:06.474 15:18:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:06.474 15:18:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:06.474 15:18:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:06.474 15:18:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:06.474 15:18:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:06.474 15:18:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:06.474 15:18:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:06.474 15:18:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:06.474 105.29 IOPS, 315.86 MiB/s [2024-11-19T15:18:56.813Z] 15:18:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:06.734 15:18:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:06.734 "name": "raid_bdev1", 00:13:06.734 "uuid": "bad7764f-d0c1-4239-9c77-e8bbb5344d80", 00:13:06.734 "strip_size_kb": 0, 00:13:06.734 "state": "online", 00:13:06.734 "raid_level": "raid1", 00:13:06.734 "superblock": false, 00:13:06.734 "num_base_bdevs": 2, 00:13:06.734 "num_base_bdevs_discovered": 2, 00:13:06.734 "num_base_bdevs_operational": 2, 00:13:06.734 "base_bdevs_list": [ 00:13:06.734 { 00:13:06.734 "name": "spare", 00:13:06.734 "uuid": "3f934860-0301-5e89-bf66-18b3cf7c62da", 00:13:06.734 "is_configured": true, 00:13:06.734 "data_offset": 0, 00:13:06.734 "data_size": 65536 00:13:06.734 }, 00:13:06.734 { 00:13:06.734 "name": "BaseBdev2", 00:13:06.734 "uuid": "17977eca-470a-5f45-aa38-6f2649a1941e", 00:13:06.734 "is_configured": true, 00:13:06.734 "data_offset": 0, 00:13:06.734 "data_size": 65536 00:13:06.734 } 00:13:06.734 ] 00:13:06.734 }' 00:13:06.734 15:18:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:06.734 15:18:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:06.734 15:18:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:06.734 15:18:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:06.734 15:18:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:06.734 15:18:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:06.734 15:18:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:06.734 15:18:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:06.734 15:18:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:06.734 15:18:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:06.734 15:18:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:06.734 15:18:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:06.734 15:18:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:06.734 15:18:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:06.734 15:18:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:06.734 15:18:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:06.734 15:18:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:06.735 15:18:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:06.735 15:18:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:06.735 15:18:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:06.735 "name": "raid_bdev1", 00:13:06.735 "uuid": "bad7764f-d0c1-4239-9c77-e8bbb5344d80", 00:13:06.735 "strip_size_kb": 0, 00:13:06.735 "state": "online", 00:13:06.735 "raid_level": "raid1", 00:13:06.735 "superblock": false, 00:13:06.735 "num_base_bdevs": 2, 00:13:06.735 "num_base_bdevs_discovered": 2, 00:13:06.735 "num_base_bdevs_operational": 2, 00:13:06.735 "base_bdevs_list": [ 00:13:06.735 { 00:13:06.735 "name": "spare", 00:13:06.735 "uuid": "3f934860-0301-5e89-bf66-18b3cf7c62da", 00:13:06.735 "is_configured": true, 00:13:06.735 "data_offset": 0, 00:13:06.735 "data_size": 65536 00:13:06.735 }, 00:13:06.735 { 00:13:06.735 "name": "BaseBdev2", 00:13:06.735 "uuid": "17977eca-470a-5f45-aa38-6f2649a1941e", 00:13:06.735 "is_configured": true, 00:13:06.735 "data_offset": 0, 00:13:06.735 "data_size": 65536 00:13:06.735 } 00:13:06.735 ] 00:13:06.735 }' 00:13:06.735 15:18:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:06.735 15:18:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:06.994 15:18:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:13:06.994 15:18:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:06.994 15:18:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:06.994 [2024-11-19 15:18:57.316704] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:06.994 [2024-11-19 15:18:57.316739] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:07.255 00:13:07.255 Latency(us) 00:13:07.255 [2024-11-19T15:18:57.594Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:07.255 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:13:07.255 raid_bdev1 : 7.60 99.39 298.16 0.00 0.00 13870.95 277.24 112183.90 00:13:07.255 [2024-11-19T15:18:57.594Z] =================================================================================================================== 00:13:07.255 [2024-11-19T15:18:57.594Z] Total : 99.39 298.16 0.00 0.00 13870.95 277.24 112183.90 00:13:07.255 [2024-11-19 15:18:57.391592] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:07.255 [2024-11-19 15:18:57.391663] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:07.255 [2024-11-19 15:18:57.391766] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:07.255 [2024-11-19 15:18:57.391811] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:13:07.255 { 00:13:07.255 "results": [ 00:13:07.255 { 00:13:07.255 "job": "raid_bdev1", 00:13:07.255 "core_mask": "0x1", 00:13:07.255 "workload": "randrw", 00:13:07.255 "percentage": 50, 00:13:07.255 "status": "finished", 00:13:07.255 "queue_depth": 2, 00:13:07.255 "io_size": 3145728, 00:13:07.255 "runtime": 7.596696, 00:13:07.255 "iops": 99.38531171972657, 00:13:07.255 "mibps": 298.1559351591797, 00:13:07.255 "io_failed": 0, 00:13:07.255 "io_timeout": 0, 00:13:07.255 "avg_latency_us": 13870.950780531537, 00:13:07.255 "min_latency_us": 277.2401746724891, 00:13:07.255 "max_latency_us": 112183.89519650655 00:13:07.255 } 00:13:07.255 ], 00:13:07.255 "core_count": 1 00:13:07.255 } 00:13:07.255 15:18:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:07.255 15:18:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:07.255 15:18:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:07.255 15:18:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:07.255 15:18:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # jq length 00:13:07.255 15:18:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:07.255 15:18:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:13:07.255 15:18:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:13:07.255 15:18:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:13:07.255 15:18:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:13:07.255 15:18:57 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:07.255 15:18:57 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:13:07.255 15:18:57 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:07.255 15:18:57 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:13:07.255 15:18:57 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:07.255 15:18:57 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:13:07.255 15:18:57 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:07.255 15:18:57 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:07.255 15:18:57 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:13:07.515 /dev/nbd0 00:13:07.515 15:18:57 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:13:07.515 15:18:57 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:13:07.515 15:18:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:13:07.515 15:18:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # local i 00:13:07.515 15:18:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:07.515 15:18:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:07.515 15:18:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:13:07.515 15:18:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@877 -- # break 00:13:07.515 15:18:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:07.515 15:18:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:07.515 15:18:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:07.515 1+0 records in 00:13:07.515 1+0 records out 00:13:07.515 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000210743 s, 19.4 MB/s 00:13:07.515 15:18:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:07.515 15:18:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # size=4096 00:13:07.515 15:18:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:07.515 15:18:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:07.515 15:18:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@893 -- # return 0 00:13:07.515 15:18:57 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:07.515 15:18:57 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:07.515 15:18:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:13:07.515 15:18:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev2 ']' 00:13:07.515 15:18:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev2 /dev/nbd1 00:13:07.515 15:18:57 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:07.515 15:18:57 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev2') 00:13:07.515 15:18:57 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:07.515 15:18:57 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:13:07.515 15:18:57 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:07.515 15:18:57 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:13:07.515 15:18:57 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:07.515 15:18:57 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:07.515 15:18:57 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev2 /dev/nbd1 00:13:07.775 /dev/nbd1 00:13:07.775 15:18:57 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:13:07.775 15:18:57 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:13:07.775 15:18:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:13:07.775 15:18:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # local i 00:13:07.775 15:18:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:07.775 15:18:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:07.775 15:18:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:13:07.775 15:18:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@877 -- # break 00:13:07.775 15:18:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:07.775 15:18:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:07.775 15:18:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:07.775 1+0 records in 00:13:07.775 1+0 records out 00:13:07.775 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00045966 s, 8.9 MB/s 00:13:07.775 15:18:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:07.775 15:18:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # size=4096 00:13:07.775 15:18:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:07.775 15:18:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:07.775 15:18:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@893 -- # return 0 00:13:07.775 15:18:57 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:07.775 15:18:57 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:07.775 15:18:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@731 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:13:07.775 15:18:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:13:07.775 15:18:57 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:07.775 15:18:57 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:13:07.775 15:18:57 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:07.775 15:18:57 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:13:07.775 15:18:57 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:07.775 15:18:57 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:13:08.035 15:18:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:13:08.035 15:18:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:13:08.035 15:18:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:13:08.035 15:18:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:08.035 15:18:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:08.035 15:18:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:13:08.035 15:18:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:13:08.035 15:18:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:13:08.035 15:18:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:13:08.035 15:18:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:08.035 15:18:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:13:08.035 15:18:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:08.035 15:18:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:13:08.035 15:18:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:08.035 15:18:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:13:08.295 15:18:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:13:08.295 15:18:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:13:08.295 15:18:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:13:08.295 15:18:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:08.295 15:18:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:08.295 15:18:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:13:08.295 15:18:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:13:08.295 15:18:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:13:08.295 15:18:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:13:08.295 15:18:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@784 -- # killprocess 87176 00:13:08.295 15:18:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@954 -- # '[' -z 87176 ']' 00:13:08.295 15:18:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@958 -- # kill -0 87176 00:13:08.295 15:18:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@959 -- # uname 00:13:08.295 15:18:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:13:08.295 15:18:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 87176 00:13:08.295 15:18:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:13:08.295 15:18:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:13:08.295 15:18:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@972 -- # echo 'killing process with pid 87176' 00:13:08.295 killing process with pid 87176 00:13:08.295 15:18:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@973 -- # kill 87176 00:13:08.295 Received shutdown signal, test time was about 8.674812 seconds 00:13:08.295 00:13:08.295 Latency(us) 00:13:08.295 [2024-11-19T15:18:58.634Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:08.295 [2024-11-19T15:18:58.634Z] =================================================================================================================== 00:13:08.295 [2024-11-19T15:18:58.634Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:13:08.295 [2024-11-19 15:18:58.464591] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:08.295 15:18:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@978 -- # wait 87176 00:13:08.295 [2024-11-19 15:18:58.490505] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:08.556 15:18:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@786 -- # return 0 00:13:08.556 00:13:08.556 real 0m10.480s 00:13:08.556 user 0m13.487s 00:13:08.556 sys 0m1.368s 00:13:08.556 15:18:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1130 -- # xtrace_disable 00:13:08.556 15:18:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:08.556 ************************************ 00:13:08.556 END TEST raid_rebuild_test_io 00:13:08.556 ************************************ 00:13:08.556 15:18:58 bdev_raid -- bdev/bdev_raid.sh@981 -- # run_test raid_rebuild_test_sb_io raid_rebuild_test raid1 2 true true true 00:13:08.556 15:18:58 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:13:08.556 15:18:58 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:13:08.556 15:18:58 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:08.556 ************************************ 00:13:08.556 START TEST raid_rebuild_test_sb_io 00:13:08.556 ************************************ 00:13:08.556 15:18:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 true true true 00:13:08.556 15:18:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:13:08.556 15:18:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:13:08.556 15:18:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:13:08.556 15:18:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:13:08.556 15:18:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:13:08.556 15:18:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:13:08.556 15:18:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:08.556 15:18:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:13:08.556 15:18:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:08.556 15:18:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:08.556 15:18:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:13:08.556 15:18:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:08.556 15:18:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:08.556 15:18:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:13:08.556 15:18:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:13:08.556 15:18:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:13:08.556 15:18:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:13:08.556 15:18:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:13:08.556 15:18:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:13:08.556 15:18:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:13:08.556 15:18:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:13:08.556 15:18:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:13:08.556 15:18:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:13:08.556 15:18:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:13:08.556 15:18:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@597 -- # raid_pid=87534 00:13:08.556 15:18:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:13:08.556 15:18:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 87534 00:13:08.556 15:18:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@835 -- # '[' -z 87534 ']' 00:13:08.556 15:18:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:08.556 15:18:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@840 -- # local max_retries=100 00:13:08.556 15:18:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:08.556 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:08.556 15:18:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@844 -- # xtrace_disable 00:13:08.556 15:18:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:08.556 [2024-11-19 15:18:58.855595] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:13:08.556 I/O size of 3145728 is greater than zero copy threshold (65536). 00:13:08.556 Zero copy mechanism will not be used. 00:13:08.556 [2024-11-19 15:18:58.855827] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid87534 ] 00:13:08.816 [2024-11-19 15:18:59.011288] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:08.816 [2024-11-19 15:18:59.035743] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:13:08.816 [2024-11-19 15:18:59.077458] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:08.816 [2024-11-19 15:18:59.077572] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:09.386 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:13:09.386 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@868 -- # return 0 00:13:09.386 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:09.386 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:13:09.386 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:09.386 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:09.386 BaseBdev1_malloc 00:13:09.386 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:09.386 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:13:09.386 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:09.386 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:09.386 [2024-11-19 15:18:59.691260] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:13:09.386 [2024-11-19 15:18:59.691327] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:09.386 [2024-11-19 15:18:59.691360] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:13:09.386 [2024-11-19 15:18:59.691372] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:09.386 [2024-11-19 15:18:59.693490] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:09.386 [2024-11-19 15:18:59.693573] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:13:09.386 BaseBdev1 00:13:09.386 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:09.386 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:09.386 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:13:09.386 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:09.386 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:09.386 BaseBdev2_malloc 00:13:09.386 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:09.386 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:13:09.386 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:09.386 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:09.386 [2024-11-19 15:18:59.719718] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:13:09.386 [2024-11-19 15:18:59.719764] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:09.386 [2024-11-19 15:18:59.719783] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:13:09.386 [2024-11-19 15:18:59.719791] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:09.386 [2024-11-19 15:18:59.721841] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:09.386 [2024-11-19 15:18:59.721881] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:13:09.648 BaseBdev2 00:13:09.648 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:09.648 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:13:09.648 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:09.648 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:09.648 spare_malloc 00:13:09.648 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:09.648 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:13:09.648 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:09.648 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:09.648 spare_delay 00:13:09.649 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:09.649 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:09.649 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:09.649 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:09.649 [2024-11-19 15:18:59.760261] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:09.649 [2024-11-19 15:18:59.760354] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:09.649 [2024-11-19 15:18:59.760394] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:13:09.649 [2024-11-19 15:18:59.760402] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:09.649 [2024-11-19 15:18:59.762460] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:09.649 [2024-11-19 15:18:59.762496] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:09.649 spare 00:13:09.649 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:09.649 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:13:09.649 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:09.649 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:09.649 [2024-11-19 15:18:59.772279] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:09.649 [2024-11-19 15:18:59.774090] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:09.649 [2024-11-19 15:18:59.774255] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:13:09.649 [2024-11-19 15:18:59.774268] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:13:09.649 [2024-11-19 15:18:59.774514] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:13:09.649 [2024-11-19 15:18:59.774645] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:13:09.649 [2024-11-19 15:18:59.774669] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:13:09.649 [2024-11-19 15:18:59.774778] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:09.649 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:09.649 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:09.649 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:09.649 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:09.649 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:09.649 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:09.649 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:09.649 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:09.649 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:09.649 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:09.649 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:09.649 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:09.649 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:09.649 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:09.649 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:09.649 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:09.649 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:09.649 "name": "raid_bdev1", 00:13:09.649 "uuid": "92391f31-3636-4e85-85ea-1e816e5d07b9", 00:13:09.649 "strip_size_kb": 0, 00:13:09.649 "state": "online", 00:13:09.649 "raid_level": "raid1", 00:13:09.649 "superblock": true, 00:13:09.649 "num_base_bdevs": 2, 00:13:09.649 "num_base_bdevs_discovered": 2, 00:13:09.649 "num_base_bdevs_operational": 2, 00:13:09.649 "base_bdevs_list": [ 00:13:09.649 { 00:13:09.649 "name": "BaseBdev1", 00:13:09.649 "uuid": "3fd046fc-6222-5e5e-8672-e8e291af4c88", 00:13:09.649 "is_configured": true, 00:13:09.649 "data_offset": 2048, 00:13:09.649 "data_size": 63488 00:13:09.649 }, 00:13:09.649 { 00:13:09.649 "name": "BaseBdev2", 00:13:09.649 "uuid": "84577df3-ff5e-5520-a176-05656fbbc918", 00:13:09.649 "is_configured": true, 00:13:09.649 "data_offset": 2048, 00:13:09.649 "data_size": 63488 00:13:09.649 } 00:13:09.649 ] 00:13:09.649 }' 00:13:09.649 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:09.649 15:18:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:09.911 15:19:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:13:09.911 15:19:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:09.911 15:19:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:09.911 15:19:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:09.911 [2024-11-19 15:19:00.232055] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:10.172 15:19:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:10.172 15:19:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:13:10.172 15:19:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:10.172 15:19:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:13:10.172 15:19:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:10.172 15:19:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:10.172 15:19:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:10.172 15:19:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:13:10.172 15:19:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:13:10.172 15:19:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:13:10.172 15:19:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:13:10.172 15:19:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:10.172 15:19:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:10.172 [2024-11-19 15:19:00.327708] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:10.172 15:19:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:10.172 15:19:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:13:10.172 15:19:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:10.172 15:19:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:10.172 15:19:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:10.172 15:19:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:10.172 15:19:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:13:10.172 15:19:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:10.172 15:19:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:10.172 15:19:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:10.172 15:19:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:10.172 15:19:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:10.172 15:19:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:10.172 15:19:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:10.172 15:19:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:10.172 15:19:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:10.172 15:19:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:10.172 "name": "raid_bdev1", 00:13:10.172 "uuid": "92391f31-3636-4e85-85ea-1e816e5d07b9", 00:13:10.172 "strip_size_kb": 0, 00:13:10.172 "state": "online", 00:13:10.172 "raid_level": "raid1", 00:13:10.172 "superblock": true, 00:13:10.172 "num_base_bdevs": 2, 00:13:10.172 "num_base_bdevs_discovered": 1, 00:13:10.172 "num_base_bdevs_operational": 1, 00:13:10.172 "base_bdevs_list": [ 00:13:10.172 { 00:13:10.172 "name": null, 00:13:10.172 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:10.172 "is_configured": false, 00:13:10.172 "data_offset": 0, 00:13:10.172 "data_size": 63488 00:13:10.172 }, 00:13:10.172 { 00:13:10.172 "name": "BaseBdev2", 00:13:10.172 "uuid": "84577df3-ff5e-5520-a176-05656fbbc918", 00:13:10.172 "is_configured": true, 00:13:10.172 "data_offset": 2048, 00:13:10.172 "data_size": 63488 00:13:10.172 } 00:13:10.172 ] 00:13:10.172 }' 00:13:10.172 15:19:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:10.172 15:19:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:10.172 [2024-11-19 15:19:00.393577] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:13:10.172 I/O size of 3145728 is greater than zero copy threshold (65536). 00:13:10.172 Zero copy mechanism will not be used. 00:13:10.172 Running I/O for 60 seconds... 00:13:10.431 15:19:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:10.431 15:19:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:10.431 15:19:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:10.431 [2024-11-19 15:19:00.763119] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:10.691 15:19:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:10.691 15:19:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:13:10.691 [2024-11-19 15:19:00.821423] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:13:10.691 [2024-11-19 15:19:00.823417] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:10.691 [2024-11-19 15:19:00.930941] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:13:10.691 [2024-11-19 15:19:00.931368] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:13:10.950 [2024-11-19 15:19:01.051070] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:13:10.950 [2024-11-19 15:19:01.051387] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:13:10.950 [2024-11-19 15:19:01.272211] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:13:10.950 [2024-11-19 15:19:01.272702] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:13:11.209 202.00 IOPS, 606.00 MiB/s [2024-11-19T15:19:01.548Z] [2024-11-19 15:19:01.490075] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:13:11.209 [2024-11-19 15:19:01.490295] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:13:11.468 15:19:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:11.468 15:19:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:11.468 15:19:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:11.468 15:19:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:11.468 15:19:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:11.727 15:19:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:11.727 15:19:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:11.727 15:19:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:11.727 [2024-11-19 15:19:01.810731] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:13:11.727 15:19:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:11.727 [2024-11-19 15:19:01.811067] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:13:11.727 15:19:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:11.727 15:19:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:11.727 "name": "raid_bdev1", 00:13:11.727 "uuid": "92391f31-3636-4e85-85ea-1e816e5d07b9", 00:13:11.727 "strip_size_kb": 0, 00:13:11.727 "state": "online", 00:13:11.727 "raid_level": "raid1", 00:13:11.727 "superblock": true, 00:13:11.727 "num_base_bdevs": 2, 00:13:11.727 "num_base_bdevs_discovered": 2, 00:13:11.727 "num_base_bdevs_operational": 2, 00:13:11.727 "process": { 00:13:11.727 "type": "rebuild", 00:13:11.727 "target": "spare", 00:13:11.727 "progress": { 00:13:11.727 "blocks": 14336, 00:13:11.727 "percent": 22 00:13:11.727 } 00:13:11.727 }, 00:13:11.727 "base_bdevs_list": [ 00:13:11.727 { 00:13:11.727 "name": "spare", 00:13:11.727 "uuid": "0c1b1169-6e12-52d9-9498-5ef8eac00e82", 00:13:11.727 "is_configured": true, 00:13:11.727 "data_offset": 2048, 00:13:11.727 "data_size": 63488 00:13:11.727 }, 00:13:11.727 { 00:13:11.727 "name": "BaseBdev2", 00:13:11.727 "uuid": "84577df3-ff5e-5520-a176-05656fbbc918", 00:13:11.727 "is_configured": true, 00:13:11.727 "data_offset": 2048, 00:13:11.727 "data_size": 63488 00:13:11.727 } 00:13:11.727 ] 00:13:11.727 }' 00:13:11.727 15:19:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:11.727 15:19:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:11.727 15:19:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:11.727 [2024-11-19 15:19:01.923721] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:13:11.727 [2024-11-19 15:19:01.923941] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:13:11.727 15:19:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:11.727 15:19:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:13:11.727 15:19:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:11.727 15:19:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:11.727 [2024-11-19 15:19:01.943813] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:11.727 [2024-11-19 15:19:02.032709] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:13:11.727 [2024-11-19 15:19:02.044824] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:11.727 [2024-11-19 15:19:02.051921] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:11.727 [2024-11-19 15:19:02.051954] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:11.727 [2024-11-19 15:19:02.051967] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:11.727 [2024-11-19 15:19:02.064343] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d0000026d0 00:13:11.986 15:19:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:11.986 15:19:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:13:11.986 15:19:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:11.986 15:19:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:11.987 15:19:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:11.987 15:19:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:11.987 15:19:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:13:11.987 15:19:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:11.987 15:19:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:11.987 15:19:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:11.987 15:19:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:11.987 15:19:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:11.987 15:19:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:11.987 15:19:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:11.987 15:19:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:11.987 15:19:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:11.987 15:19:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:11.987 "name": "raid_bdev1", 00:13:11.987 "uuid": "92391f31-3636-4e85-85ea-1e816e5d07b9", 00:13:11.987 "strip_size_kb": 0, 00:13:11.987 "state": "online", 00:13:11.987 "raid_level": "raid1", 00:13:11.987 "superblock": true, 00:13:11.987 "num_base_bdevs": 2, 00:13:11.987 "num_base_bdevs_discovered": 1, 00:13:11.987 "num_base_bdevs_operational": 1, 00:13:11.987 "base_bdevs_list": [ 00:13:11.987 { 00:13:11.987 "name": null, 00:13:11.987 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:11.987 "is_configured": false, 00:13:11.987 "data_offset": 0, 00:13:11.987 "data_size": 63488 00:13:11.987 }, 00:13:11.987 { 00:13:11.987 "name": "BaseBdev2", 00:13:11.987 "uuid": "84577df3-ff5e-5520-a176-05656fbbc918", 00:13:11.987 "is_configured": true, 00:13:11.987 "data_offset": 2048, 00:13:11.987 "data_size": 63488 00:13:11.987 } 00:13:11.987 ] 00:13:11.987 }' 00:13:11.987 15:19:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:11.987 15:19:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:12.246 173.00 IOPS, 519.00 MiB/s [2024-11-19T15:19:02.585Z] 15:19:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:12.246 15:19:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:12.246 15:19:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:12.246 15:19:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:12.246 15:19:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:12.246 15:19:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:12.246 15:19:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:12.246 15:19:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:12.246 15:19:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:12.246 15:19:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:12.246 15:19:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:12.246 "name": "raid_bdev1", 00:13:12.246 "uuid": "92391f31-3636-4e85-85ea-1e816e5d07b9", 00:13:12.246 "strip_size_kb": 0, 00:13:12.246 "state": "online", 00:13:12.246 "raid_level": "raid1", 00:13:12.246 "superblock": true, 00:13:12.246 "num_base_bdevs": 2, 00:13:12.246 "num_base_bdevs_discovered": 1, 00:13:12.246 "num_base_bdevs_operational": 1, 00:13:12.246 "base_bdevs_list": [ 00:13:12.246 { 00:13:12.246 "name": null, 00:13:12.246 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:12.246 "is_configured": false, 00:13:12.246 "data_offset": 0, 00:13:12.246 "data_size": 63488 00:13:12.246 }, 00:13:12.246 { 00:13:12.246 "name": "BaseBdev2", 00:13:12.246 "uuid": "84577df3-ff5e-5520-a176-05656fbbc918", 00:13:12.246 "is_configured": true, 00:13:12.246 "data_offset": 2048, 00:13:12.246 "data_size": 63488 00:13:12.246 } 00:13:12.246 ] 00:13:12.246 }' 00:13:12.246 15:19:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:12.246 15:19:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:12.246 15:19:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:12.504 15:19:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:12.504 15:19:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:12.504 15:19:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:12.504 15:19:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:12.504 [2024-11-19 15:19:02.629486] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:12.504 15:19:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:12.504 15:19:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:13:12.504 [2024-11-19 15:19:02.661274] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:13:12.504 [2024-11-19 15:19:02.663204] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:12.504 [2024-11-19 15:19:02.775491] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:13:12.504 [2024-11-19 15:19:02.776032] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:13:12.763 [2024-11-19 15:19:02.983830] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:13:12.763 [2024-11-19 15:19:02.984079] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:13:13.021 [2024-11-19 15:19:03.210044] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:13:13.021 [2024-11-19 15:19:03.322876] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:13:13.021 [2024-11-19 15:19:03.323094] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:13:13.556 177.33 IOPS, 532.00 MiB/s [2024-11-19T15:19:03.895Z] 15:19:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:13.556 15:19:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:13.556 15:19:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:13.556 15:19:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:13.556 15:19:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:13.556 15:19:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:13.556 15:19:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:13.556 15:19:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:13.556 15:19:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:13.556 [2024-11-19 15:19:03.666602] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:13:13.556 15:19:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:13.556 15:19:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:13.556 "name": "raid_bdev1", 00:13:13.556 "uuid": "92391f31-3636-4e85-85ea-1e816e5d07b9", 00:13:13.556 "strip_size_kb": 0, 00:13:13.556 "state": "online", 00:13:13.556 "raid_level": "raid1", 00:13:13.556 "superblock": true, 00:13:13.556 "num_base_bdevs": 2, 00:13:13.556 "num_base_bdevs_discovered": 2, 00:13:13.556 "num_base_bdevs_operational": 2, 00:13:13.556 "process": { 00:13:13.556 "type": "rebuild", 00:13:13.556 "target": "spare", 00:13:13.556 "progress": { 00:13:13.556 "blocks": 12288, 00:13:13.556 "percent": 19 00:13:13.556 } 00:13:13.556 }, 00:13:13.556 "base_bdevs_list": [ 00:13:13.556 { 00:13:13.556 "name": "spare", 00:13:13.556 "uuid": "0c1b1169-6e12-52d9-9498-5ef8eac00e82", 00:13:13.556 "is_configured": true, 00:13:13.556 "data_offset": 2048, 00:13:13.556 "data_size": 63488 00:13:13.556 }, 00:13:13.556 { 00:13:13.556 "name": "BaseBdev2", 00:13:13.556 "uuid": "84577df3-ff5e-5520-a176-05656fbbc918", 00:13:13.556 "is_configured": true, 00:13:13.556 "data_offset": 2048, 00:13:13.556 "data_size": 63488 00:13:13.556 } 00:13:13.556 ] 00:13:13.556 }' 00:13:13.556 15:19:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:13.556 15:19:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:13.556 15:19:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:13.556 [2024-11-19 15:19:03.774687] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:13:13.556 15:19:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:13.556 15:19:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:13:13.557 15:19:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:13:13.557 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:13:13.557 15:19:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:13:13.557 15:19:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:13:13.557 15:19:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:13:13.557 15:19:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@706 -- # local timeout=341 00:13:13.557 15:19:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:13.557 15:19:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:13.557 15:19:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:13.557 15:19:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:13.557 15:19:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:13.557 15:19:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:13.557 15:19:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:13.557 15:19:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:13.557 15:19:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:13.557 15:19:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:13.557 15:19:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:13.557 15:19:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:13.557 "name": "raid_bdev1", 00:13:13.557 "uuid": "92391f31-3636-4e85-85ea-1e816e5d07b9", 00:13:13.557 "strip_size_kb": 0, 00:13:13.557 "state": "online", 00:13:13.557 "raid_level": "raid1", 00:13:13.557 "superblock": true, 00:13:13.557 "num_base_bdevs": 2, 00:13:13.557 "num_base_bdevs_discovered": 2, 00:13:13.557 "num_base_bdevs_operational": 2, 00:13:13.557 "process": { 00:13:13.557 "type": "rebuild", 00:13:13.557 "target": "spare", 00:13:13.557 "progress": { 00:13:13.557 "blocks": 16384, 00:13:13.557 "percent": 25 00:13:13.557 } 00:13:13.557 }, 00:13:13.557 "base_bdevs_list": [ 00:13:13.557 { 00:13:13.557 "name": "spare", 00:13:13.557 "uuid": "0c1b1169-6e12-52d9-9498-5ef8eac00e82", 00:13:13.557 "is_configured": true, 00:13:13.557 "data_offset": 2048, 00:13:13.557 "data_size": 63488 00:13:13.557 }, 00:13:13.557 { 00:13:13.557 "name": "BaseBdev2", 00:13:13.557 "uuid": "84577df3-ff5e-5520-a176-05656fbbc918", 00:13:13.557 "is_configured": true, 00:13:13.557 "data_offset": 2048, 00:13:13.557 "data_size": 63488 00:13:13.557 } 00:13:13.557 ] 00:13:13.557 }' 00:13:13.557 15:19:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:13.557 15:19:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:13.557 15:19:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:13.842 15:19:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:13.842 15:19:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:13.842 [2024-11-19 15:19:04.101591] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:13:14.101 [2024-11-19 15:19:04.220011] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:13:14.671 150.25 IOPS, 450.75 MiB/s [2024-11-19T15:19:05.010Z] [2024-11-19 15:19:04.870409] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 32768 offset_begin: 30720 offset_end: 36864 00:13:14.671 15:19:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:14.671 15:19:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:14.671 15:19:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:14.671 15:19:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:14.671 15:19:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:14.671 15:19:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:14.671 15:19:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:14.671 15:19:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:14.671 15:19:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:14.671 15:19:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:14.671 15:19:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:14.671 15:19:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:14.671 "name": "raid_bdev1", 00:13:14.671 "uuid": "92391f31-3636-4e85-85ea-1e816e5d07b9", 00:13:14.671 "strip_size_kb": 0, 00:13:14.671 "state": "online", 00:13:14.671 "raid_level": "raid1", 00:13:14.671 "superblock": true, 00:13:14.671 "num_base_bdevs": 2, 00:13:14.671 "num_base_bdevs_discovered": 2, 00:13:14.671 "num_base_bdevs_operational": 2, 00:13:14.671 "process": { 00:13:14.671 "type": "rebuild", 00:13:14.671 "target": "spare", 00:13:14.672 "progress": { 00:13:14.672 "blocks": 32768, 00:13:14.672 "percent": 51 00:13:14.672 } 00:13:14.672 }, 00:13:14.672 "base_bdevs_list": [ 00:13:14.672 { 00:13:14.672 "name": "spare", 00:13:14.672 "uuid": "0c1b1169-6e12-52d9-9498-5ef8eac00e82", 00:13:14.672 "is_configured": true, 00:13:14.672 "data_offset": 2048, 00:13:14.672 "data_size": 63488 00:13:14.672 }, 00:13:14.672 { 00:13:14.672 "name": "BaseBdev2", 00:13:14.672 "uuid": "84577df3-ff5e-5520-a176-05656fbbc918", 00:13:14.672 "is_configured": true, 00:13:14.672 "data_offset": 2048, 00:13:14.672 "data_size": 63488 00:13:14.672 } 00:13:14.672 ] 00:13:14.672 }' 00:13:14.672 15:19:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:14.672 [2024-11-19 15:19:04.977330] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 34816 offset_begin: 30720 offset_end: 36864 00:13:14.672 15:19:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:14.672 15:19:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:14.931 15:19:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:14.931 15:19:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:14.931 [2024-11-19 15:19:05.185428] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 38912 offset_begin: 36864 offset_end: 43008 00:13:15.451 133.20 IOPS, 399.60 MiB/s [2024-11-19T15:19:05.790Z] [2024-11-19 15:19:05.624429] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 47104 offset_begin: 43008 offset_end: 49152 00:13:15.451 [2024-11-19 15:19:05.624652] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 47104 offset_begin: 43008 offset_end: 49152 00:13:16.020 15:19:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:16.020 15:19:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:16.020 15:19:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:16.020 15:19:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:16.020 15:19:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:16.020 15:19:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:16.020 [2024-11-19 15:19:06.062434] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 53248 offset_begin: 49152 offset_end: 55296 00:13:16.020 15:19:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:16.020 15:19:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:16.020 15:19:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:16.020 15:19:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:16.020 15:19:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:16.020 15:19:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:16.020 "name": "raid_bdev1", 00:13:16.020 "uuid": "92391f31-3636-4e85-85ea-1e816e5d07b9", 00:13:16.020 "strip_size_kb": 0, 00:13:16.020 "state": "online", 00:13:16.020 "raid_level": "raid1", 00:13:16.020 "superblock": true, 00:13:16.020 "num_base_bdevs": 2, 00:13:16.020 "num_base_bdevs_discovered": 2, 00:13:16.020 "num_base_bdevs_operational": 2, 00:13:16.020 "process": { 00:13:16.020 "type": "rebuild", 00:13:16.020 "target": "spare", 00:13:16.020 "progress": { 00:13:16.020 "blocks": 53248, 00:13:16.020 "percent": 83 00:13:16.020 } 00:13:16.020 }, 00:13:16.020 "base_bdevs_list": [ 00:13:16.020 { 00:13:16.020 "name": "spare", 00:13:16.020 "uuid": "0c1b1169-6e12-52d9-9498-5ef8eac00e82", 00:13:16.020 "is_configured": true, 00:13:16.020 "data_offset": 2048, 00:13:16.020 "data_size": 63488 00:13:16.020 }, 00:13:16.020 { 00:13:16.020 "name": "BaseBdev2", 00:13:16.020 "uuid": "84577df3-ff5e-5520-a176-05656fbbc918", 00:13:16.020 "is_configured": true, 00:13:16.020 "data_offset": 2048, 00:13:16.020 "data_size": 63488 00:13:16.020 } 00:13:16.020 ] 00:13:16.020 }' 00:13:16.020 15:19:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:16.020 15:19:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:16.020 15:19:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:16.020 15:19:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:16.020 15:19:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:16.020 [2024-11-19 15:19:06.287662] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 57344 offset_begin: 55296 offset_end: 61440 00:13:16.539 117.33 IOPS, 352.00 MiB/s [2024-11-19T15:19:06.878Z] [2024-11-19 15:19:06.715354] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:13:16.539 [2024-11-19 15:19:06.815223] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:13:16.539 [2024-11-19 15:19:06.816438] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:17.108 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:17.108 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:17.108 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:17.108 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:17.108 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:17.108 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:17.108 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:17.108 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:17.108 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:17.108 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:17.108 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:17.108 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:17.108 "name": "raid_bdev1", 00:13:17.108 "uuid": "92391f31-3636-4e85-85ea-1e816e5d07b9", 00:13:17.108 "strip_size_kb": 0, 00:13:17.108 "state": "online", 00:13:17.108 "raid_level": "raid1", 00:13:17.108 "superblock": true, 00:13:17.108 "num_base_bdevs": 2, 00:13:17.108 "num_base_bdevs_discovered": 2, 00:13:17.108 "num_base_bdevs_operational": 2, 00:13:17.108 "base_bdevs_list": [ 00:13:17.108 { 00:13:17.108 "name": "spare", 00:13:17.108 "uuid": "0c1b1169-6e12-52d9-9498-5ef8eac00e82", 00:13:17.108 "is_configured": true, 00:13:17.108 "data_offset": 2048, 00:13:17.108 "data_size": 63488 00:13:17.108 }, 00:13:17.108 { 00:13:17.108 "name": "BaseBdev2", 00:13:17.108 "uuid": "84577df3-ff5e-5520-a176-05656fbbc918", 00:13:17.108 "is_configured": true, 00:13:17.108 "data_offset": 2048, 00:13:17.108 "data_size": 63488 00:13:17.108 } 00:13:17.108 ] 00:13:17.108 }' 00:13:17.108 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:17.108 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:13:17.108 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:17.108 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:13:17.108 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@709 -- # break 00:13:17.108 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:17.108 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:17.108 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:17.108 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:17.109 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:17.109 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:17.109 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:17.109 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:17.109 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:17.109 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:17.109 106.14 IOPS, 318.43 MiB/s [2024-11-19T15:19:07.448Z] 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:17.109 "name": "raid_bdev1", 00:13:17.109 "uuid": "92391f31-3636-4e85-85ea-1e816e5d07b9", 00:13:17.109 "strip_size_kb": 0, 00:13:17.109 "state": "online", 00:13:17.109 "raid_level": "raid1", 00:13:17.109 "superblock": true, 00:13:17.109 "num_base_bdevs": 2, 00:13:17.109 "num_base_bdevs_discovered": 2, 00:13:17.109 "num_base_bdevs_operational": 2, 00:13:17.109 "base_bdevs_list": [ 00:13:17.109 { 00:13:17.109 "name": "spare", 00:13:17.109 "uuid": "0c1b1169-6e12-52d9-9498-5ef8eac00e82", 00:13:17.109 "is_configured": true, 00:13:17.109 "data_offset": 2048, 00:13:17.109 "data_size": 63488 00:13:17.109 }, 00:13:17.109 { 00:13:17.109 "name": "BaseBdev2", 00:13:17.109 "uuid": "84577df3-ff5e-5520-a176-05656fbbc918", 00:13:17.109 "is_configured": true, 00:13:17.109 "data_offset": 2048, 00:13:17.109 "data_size": 63488 00:13:17.109 } 00:13:17.109 ] 00:13:17.109 }' 00:13:17.109 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:17.368 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:17.368 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:17.368 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:17.368 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:17.368 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:17.368 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:17.368 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:17.368 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:17.368 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:17.368 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:17.368 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:17.368 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:17.368 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:17.368 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:17.368 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:17.368 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:17.368 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:17.368 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:17.368 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:17.368 "name": "raid_bdev1", 00:13:17.368 "uuid": "92391f31-3636-4e85-85ea-1e816e5d07b9", 00:13:17.368 "strip_size_kb": 0, 00:13:17.368 "state": "online", 00:13:17.368 "raid_level": "raid1", 00:13:17.368 "superblock": true, 00:13:17.368 "num_base_bdevs": 2, 00:13:17.368 "num_base_bdevs_discovered": 2, 00:13:17.368 "num_base_bdevs_operational": 2, 00:13:17.368 "base_bdevs_list": [ 00:13:17.368 { 00:13:17.368 "name": "spare", 00:13:17.368 "uuid": "0c1b1169-6e12-52d9-9498-5ef8eac00e82", 00:13:17.368 "is_configured": true, 00:13:17.368 "data_offset": 2048, 00:13:17.368 "data_size": 63488 00:13:17.368 }, 00:13:17.368 { 00:13:17.368 "name": "BaseBdev2", 00:13:17.368 "uuid": "84577df3-ff5e-5520-a176-05656fbbc918", 00:13:17.368 "is_configured": true, 00:13:17.368 "data_offset": 2048, 00:13:17.368 "data_size": 63488 00:13:17.368 } 00:13:17.368 ] 00:13:17.368 }' 00:13:17.368 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:17.368 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:17.626 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:13:17.626 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:17.626 15:19:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:17.626 [2024-11-19 15:19:07.932844] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:17.626 [2024-11-19 15:19:07.932937] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:17.885 00:13:17.885 Latency(us) 00:13:17.885 [2024-11-19T15:19:08.224Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:17.885 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:13:17.885 raid_bdev1 : 7.63 100.69 302.06 0.00 0.00 13212.83 275.45 107604.96 00:13:17.885 [2024-11-19T15:19:08.224Z] =================================================================================================================== 00:13:17.885 [2024-11-19T15:19:08.224Z] Total : 100.69 302.06 0.00 0.00 13212.83 275.45 107604.96 00:13:17.886 [2024-11-19 15:19:08.011727] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:17.886 [2024-11-19 15:19:08.011797] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:17.886 [2024-11-19 15:19:08.011890] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:17.886 [2024-11-19 15:19:08.011942] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:13:17.886 { 00:13:17.886 "results": [ 00:13:17.886 { 00:13:17.886 "job": "raid_bdev1", 00:13:17.886 "core_mask": "0x1", 00:13:17.886 "workload": "randrw", 00:13:17.886 "percentage": 50, 00:13:17.886 "status": "finished", 00:13:17.886 "queue_depth": 2, 00:13:17.886 "io_size": 3145728, 00:13:17.886 "runtime": 7.627616, 00:13:17.886 "iops": 100.68676766108834, 00:13:17.886 "mibps": 302.06030298326505, 00:13:17.886 "io_failed": 0, 00:13:17.886 "io_timeout": 0, 00:13:17.886 "avg_latency_us": 13212.833770014555, 00:13:17.886 "min_latency_us": 275.45152838427947, 00:13:17.886 "max_latency_us": 107604.96069868996 00:13:17.886 } 00:13:17.886 ], 00:13:17.886 "core_count": 1 00:13:17.886 } 00:13:17.886 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:17.886 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:17.886 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:17.886 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:17.886 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # jq length 00:13:17.886 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:17.886 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:13:17.886 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:13:17.886 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:13:17.886 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:13:17.886 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:17.886 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:13:17.886 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:17.886 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:13:17.886 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:17.886 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:13:17.886 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:17.886 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:17.886 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:13:18.145 /dev/nbd0 00:13:18.146 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:13:18.146 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:13:18.146 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:13:18.146 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # local i 00:13:18.146 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:18.146 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:18.146 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:13:18.146 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@877 -- # break 00:13:18.146 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:18.146 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:18.146 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:18.146 1+0 records in 00:13:18.146 1+0 records out 00:13:18.146 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000340946 s, 12.0 MB/s 00:13:18.146 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:18.146 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # size=4096 00:13:18.146 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:18.146 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:18.146 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@893 -- # return 0 00:13:18.146 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:18.146 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:18.146 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:13:18.146 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev2 ']' 00:13:18.146 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev2 /dev/nbd1 00:13:18.146 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:18.146 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev2') 00:13:18.146 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:18.146 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:13:18.146 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:18.146 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:13:18.146 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:18.146 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:18.146 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev2 /dev/nbd1 00:13:18.405 /dev/nbd1 00:13:18.405 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:13:18.405 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:13:18.405 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:13:18.405 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # local i 00:13:18.405 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:18.405 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:18.405 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:13:18.405 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@877 -- # break 00:13:18.405 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:18.405 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:18.405 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:18.405 1+0 records in 00:13:18.405 1+0 records out 00:13:18.405 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000370626 s, 11.1 MB/s 00:13:18.405 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:18.405 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # size=4096 00:13:18.405 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:18.405 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:18.405 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@893 -- # return 0 00:13:18.405 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:18.405 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:18.406 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@731 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:13:18.406 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:13:18.406 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:18.406 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:13:18.406 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:18.406 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:13:18.406 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:18.406 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:13:18.666 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:13:18.666 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:13:18.666 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:13:18.666 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:18.666 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:18.666 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:13:18.666 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:13:18.666 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:13:18.666 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:13:18.666 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:18.666 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:13:18.666 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:18.666 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:13:18.666 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:18.666 15:19:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:13:18.926 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:13:18.926 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:13:18.926 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:13:18.926 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:18.926 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:18.926 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:13:18.926 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:13:18.926 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:13:18.926 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:13:18.926 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:13:18.926 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:18.926 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:18.926 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:18.926 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:18.926 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:18.926 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:18.926 [2024-11-19 15:19:09.061444] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:18.926 [2024-11-19 15:19:09.061514] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:18.926 [2024-11-19 15:19:09.061537] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009980 00:13:18.926 [2024-11-19 15:19:09.061547] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:18.926 [2024-11-19 15:19:09.063643] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:18.926 [2024-11-19 15:19:09.063680] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:18.926 [2024-11-19 15:19:09.063791] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:13:18.926 [2024-11-19 15:19:09.063826] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:18.926 [2024-11-19 15:19:09.063979] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:18.926 spare 00:13:18.926 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:18.926 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:13:18.926 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:18.926 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:18.926 [2024-11-19 15:19:09.163875] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:13:18.926 [2024-11-19 15:19:09.163907] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:13:18.926 [2024-11-19 15:19:09.164187] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000027720 00:13:18.926 [2024-11-19 15:19:09.164338] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:13:18.926 [2024-11-19 15:19:09.164371] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:13:18.926 [2024-11-19 15:19:09.164502] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:18.926 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:18.926 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:18.926 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:18.926 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:18.926 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:18.926 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:18.926 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:18.926 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:18.926 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:18.926 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:18.926 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:18.926 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:18.926 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:18.926 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:18.926 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:18.926 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:18.926 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:18.926 "name": "raid_bdev1", 00:13:18.926 "uuid": "92391f31-3636-4e85-85ea-1e816e5d07b9", 00:13:18.926 "strip_size_kb": 0, 00:13:18.926 "state": "online", 00:13:18.926 "raid_level": "raid1", 00:13:18.926 "superblock": true, 00:13:18.926 "num_base_bdevs": 2, 00:13:18.926 "num_base_bdevs_discovered": 2, 00:13:18.926 "num_base_bdevs_operational": 2, 00:13:18.926 "base_bdevs_list": [ 00:13:18.926 { 00:13:18.926 "name": "spare", 00:13:18.926 "uuid": "0c1b1169-6e12-52d9-9498-5ef8eac00e82", 00:13:18.926 "is_configured": true, 00:13:18.926 "data_offset": 2048, 00:13:18.926 "data_size": 63488 00:13:18.926 }, 00:13:18.926 { 00:13:18.926 "name": "BaseBdev2", 00:13:18.926 "uuid": "84577df3-ff5e-5520-a176-05656fbbc918", 00:13:18.926 "is_configured": true, 00:13:18.926 "data_offset": 2048, 00:13:18.926 "data_size": 63488 00:13:18.926 } 00:13:18.926 ] 00:13:18.926 }' 00:13:18.926 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:18.926 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:19.496 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:19.496 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:19.496 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:19.496 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:19.496 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:19.496 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:19.496 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:19.496 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:19.496 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:19.496 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:19.496 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:19.496 "name": "raid_bdev1", 00:13:19.496 "uuid": "92391f31-3636-4e85-85ea-1e816e5d07b9", 00:13:19.496 "strip_size_kb": 0, 00:13:19.496 "state": "online", 00:13:19.496 "raid_level": "raid1", 00:13:19.496 "superblock": true, 00:13:19.496 "num_base_bdevs": 2, 00:13:19.496 "num_base_bdevs_discovered": 2, 00:13:19.496 "num_base_bdevs_operational": 2, 00:13:19.496 "base_bdevs_list": [ 00:13:19.496 { 00:13:19.496 "name": "spare", 00:13:19.496 "uuid": "0c1b1169-6e12-52d9-9498-5ef8eac00e82", 00:13:19.496 "is_configured": true, 00:13:19.496 "data_offset": 2048, 00:13:19.496 "data_size": 63488 00:13:19.496 }, 00:13:19.496 { 00:13:19.496 "name": "BaseBdev2", 00:13:19.496 "uuid": "84577df3-ff5e-5520-a176-05656fbbc918", 00:13:19.496 "is_configured": true, 00:13:19.496 "data_offset": 2048, 00:13:19.496 "data_size": 63488 00:13:19.496 } 00:13:19.496 ] 00:13:19.496 }' 00:13:19.496 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:19.496 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:19.496 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:19.496 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:19.496 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:19.496 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:19.496 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:19.496 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:13:19.496 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:19.496 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:13:19.496 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:13:19.496 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:19.496 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:19.496 [2024-11-19 15:19:09.748381] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:19.496 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:19.496 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:13:19.496 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:19.496 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:19.496 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:19.496 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:19.496 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:13:19.496 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:19.496 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:19.496 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:19.496 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:19.496 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:19.496 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:19.496 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:19.496 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:19.496 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:19.496 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:19.496 "name": "raid_bdev1", 00:13:19.496 "uuid": "92391f31-3636-4e85-85ea-1e816e5d07b9", 00:13:19.496 "strip_size_kb": 0, 00:13:19.496 "state": "online", 00:13:19.496 "raid_level": "raid1", 00:13:19.496 "superblock": true, 00:13:19.496 "num_base_bdevs": 2, 00:13:19.496 "num_base_bdevs_discovered": 1, 00:13:19.496 "num_base_bdevs_operational": 1, 00:13:19.496 "base_bdevs_list": [ 00:13:19.496 { 00:13:19.496 "name": null, 00:13:19.496 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:19.496 "is_configured": false, 00:13:19.496 "data_offset": 0, 00:13:19.496 "data_size": 63488 00:13:19.496 }, 00:13:19.496 { 00:13:19.496 "name": "BaseBdev2", 00:13:19.496 "uuid": "84577df3-ff5e-5520-a176-05656fbbc918", 00:13:19.496 "is_configured": true, 00:13:19.496 "data_offset": 2048, 00:13:19.496 "data_size": 63488 00:13:19.496 } 00:13:19.496 ] 00:13:19.496 }' 00:13:19.496 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:19.496 15:19:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:20.066 15:19:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:20.066 15:19:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:20.066 15:19:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:20.066 [2024-11-19 15:19:10.171827] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:20.066 [2024-11-19 15:19:10.172016] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:13:20.066 [2024-11-19 15:19:10.172035] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:13:20.066 [2024-11-19 15:19:10.172077] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:20.066 [2024-11-19 15:19:10.177228] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000277f0 00:13:20.066 15:19:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:20.066 15:19:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@757 -- # sleep 1 00:13:20.066 [2024-11-19 15:19:10.179087] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:21.007 15:19:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:21.007 15:19:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:21.007 15:19:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:21.007 15:19:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:21.007 15:19:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:21.007 15:19:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:21.007 15:19:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:21.007 15:19:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:21.007 15:19:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:21.007 15:19:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:21.007 15:19:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:21.007 "name": "raid_bdev1", 00:13:21.007 "uuid": "92391f31-3636-4e85-85ea-1e816e5d07b9", 00:13:21.007 "strip_size_kb": 0, 00:13:21.007 "state": "online", 00:13:21.007 "raid_level": "raid1", 00:13:21.007 "superblock": true, 00:13:21.007 "num_base_bdevs": 2, 00:13:21.007 "num_base_bdevs_discovered": 2, 00:13:21.007 "num_base_bdevs_operational": 2, 00:13:21.007 "process": { 00:13:21.007 "type": "rebuild", 00:13:21.007 "target": "spare", 00:13:21.007 "progress": { 00:13:21.007 "blocks": 20480, 00:13:21.007 "percent": 32 00:13:21.007 } 00:13:21.007 }, 00:13:21.007 "base_bdevs_list": [ 00:13:21.007 { 00:13:21.007 "name": "spare", 00:13:21.007 "uuid": "0c1b1169-6e12-52d9-9498-5ef8eac00e82", 00:13:21.007 "is_configured": true, 00:13:21.007 "data_offset": 2048, 00:13:21.007 "data_size": 63488 00:13:21.007 }, 00:13:21.007 { 00:13:21.007 "name": "BaseBdev2", 00:13:21.007 "uuid": "84577df3-ff5e-5520-a176-05656fbbc918", 00:13:21.007 "is_configured": true, 00:13:21.007 "data_offset": 2048, 00:13:21.007 "data_size": 63488 00:13:21.007 } 00:13:21.007 ] 00:13:21.007 }' 00:13:21.007 15:19:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:21.007 15:19:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:21.007 15:19:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:21.007 15:19:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:21.007 15:19:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:13:21.007 15:19:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:21.007 15:19:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:21.007 [2024-11-19 15:19:11.315866] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:21.267 [2024-11-19 15:19:11.382981] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:21.267 [2024-11-19 15:19:11.383039] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:21.268 [2024-11-19 15:19:11.383069] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:21.268 [2024-11-19 15:19:11.383077] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:21.268 15:19:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:21.268 15:19:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:13:21.268 15:19:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:21.268 15:19:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:21.268 15:19:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:21.268 15:19:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:21.268 15:19:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:13:21.268 15:19:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:21.268 15:19:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:21.268 15:19:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:21.268 15:19:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:21.268 15:19:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:21.268 15:19:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:21.268 15:19:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:21.268 15:19:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:21.268 15:19:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:21.268 15:19:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:21.268 "name": "raid_bdev1", 00:13:21.268 "uuid": "92391f31-3636-4e85-85ea-1e816e5d07b9", 00:13:21.268 "strip_size_kb": 0, 00:13:21.268 "state": "online", 00:13:21.268 "raid_level": "raid1", 00:13:21.268 "superblock": true, 00:13:21.268 "num_base_bdevs": 2, 00:13:21.268 "num_base_bdevs_discovered": 1, 00:13:21.268 "num_base_bdevs_operational": 1, 00:13:21.268 "base_bdevs_list": [ 00:13:21.268 { 00:13:21.268 "name": null, 00:13:21.268 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:21.268 "is_configured": false, 00:13:21.268 "data_offset": 0, 00:13:21.268 "data_size": 63488 00:13:21.268 }, 00:13:21.268 { 00:13:21.268 "name": "BaseBdev2", 00:13:21.268 "uuid": "84577df3-ff5e-5520-a176-05656fbbc918", 00:13:21.268 "is_configured": true, 00:13:21.268 "data_offset": 2048, 00:13:21.268 "data_size": 63488 00:13:21.268 } 00:13:21.268 ] 00:13:21.268 }' 00:13:21.268 15:19:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:21.268 15:19:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:21.838 15:19:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:21.838 15:19:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:21.838 15:19:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:21.838 [2024-11-19 15:19:11.894935] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:21.838 [2024-11-19 15:19:11.895003] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:21.838 [2024-11-19 15:19:11.895026] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:13:21.838 [2024-11-19 15:19:11.895037] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:21.838 [2024-11-19 15:19:11.895476] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:21.838 [2024-11-19 15:19:11.895507] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:21.838 [2024-11-19 15:19:11.895595] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:13:21.838 [2024-11-19 15:19:11.895621] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:13:21.838 [2024-11-19 15:19:11.895631] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:13:21.838 [2024-11-19 15:19:11.895672] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:21.838 [2024-11-19 15:19:11.900556] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000278c0 00:13:21.838 spare 00:13:21.838 15:19:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:21.838 15:19:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@764 -- # sleep 1 00:13:21.838 [2024-11-19 15:19:11.902518] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:22.778 15:19:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:22.778 15:19:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:22.778 15:19:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:22.778 15:19:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:22.778 15:19:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:22.778 15:19:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:22.778 15:19:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:22.778 15:19:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:22.778 15:19:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:22.778 15:19:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:22.778 15:19:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:22.778 "name": "raid_bdev1", 00:13:22.778 "uuid": "92391f31-3636-4e85-85ea-1e816e5d07b9", 00:13:22.778 "strip_size_kb": 0, 00:13:22.778 "state": "online", 00:13:22.778 "raid_level": "raid1", 00:13:22.778 "superblock": true, 00:13:22.778 "num_base_bdevs": 2, 00:13:22.778 "num_base_bdevs_discovered": 2, 00:13:22.778 "num_base_bdevs_operational": 2, 00:13:22.778 "process": { 00:13:22.778 "type": "rebuild", 00:13:22.778 "target": "spare", 00:13:22.778 "progress": { 00:13:22.778 "blocks": 20480, 00:13:22.778 "percent": 32 00:13:22.778 } 00:13:22.778 }, 00:13:22.778 "base_bdevs_list": [ 00:13:22.778 { 00:13:22.778 "name": "spare", 00:13:22.778 "uuid": "0c1b1169-6e12-52d9-9498-5ef8eac00e82", 00:13:22.778 "is_configured": true, 00:13:22.778 "data_offset": 2048, 00:13:22.778 "data_size": 63488 00:13:22.778 }, 00:13:22.778 { 00:13:22.778 "name": "BaseBdev2", 00:13:22.779 "uuid": "84577df3-ff5e-5520-a176-05656fbbc918", 00:13:22.779 "is_configured": true, 00:13:22.779 "data_offset": 2048, 00:13:22.779 "data_size": 63488 00:13:22.779 } 00:13:22.779 ] 00:13:22.779 }' 00:13:22.779 15:19:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:22.779 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:22.779 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:22.779 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:22.779 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:13:22.779 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:22.779 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:22.779 [2024-11-19 15:19:13.059895] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:22.779 [2024-11-19 15:19:13.106519] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:22.779 [2024-11-19 15:19:13.106577] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:22.779 [2024-11-19 15:19:13.106609] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:22.779 [2024-11-19 15:19:13.106616] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:23.039 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:23.039 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:13:23.039 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:23.039 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:23.039 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:23.039 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:23.039 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:13:23.039 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:23.039 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:23.039 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:23.039 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:23.039 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:23.039 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:23.039 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:23.039 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:23.039 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:23.039 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:23.039 "name": "raid_bdev1", 00:13:23.039 "uuid": "92391f31-3636-4e85-85ea-1e816e5d07b9", 00:13:23.039 "strip_size_kb": 0, 00:13:23.039 "state": "online", 00:13:23.039 "raid_level": "raid1", 00:13:23.039 "superblock": true, 00:13:23.039 "num_base_bdevs": 2, 00:13:23.039 "num_base_bdevs_discovered": 1, 00:13:23.039 "num_base_bdevs_operational": 1, 00:13:23.039 "base_bdevs_list": [ 00:13:23.039 { 00:13:23.039 "name": null, 00:13:23.039 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:23.039 "is_configured": false, 00:13:23.039 "data_offset": 0, 00:13:23.039 "data_size": 63488 00:13:23.039 }, 00:13:23.039 { 00:13:23.039 "name": "BaseBdev2", 00:13:23.039 "uuid": "84577df3-ff5e-5520-a176-05656fbbc918", 00:13:23.039 "is_configured": true, 00:13:23.039 "data_offset": 2048, 00:13:23.039 "data_size": 63488 00:13:23.039 } 00:13:23.039 ] 00:13:23.039 }' 00:13:23.039 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:23.039 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:23.299 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:23.299 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:23.299 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:23.299 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:23.299 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:23.299 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:23.299 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:23.299 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:23.299 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:23.299 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:23.299 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:23.299 "name": "raid_bdev1", 00:13:23.299 "uuid": "92391f31-3636-4e85-85ea-1e816e5d07b9", 00:13:23.299 "strip_size_kb": 0, 00:13:23.299 "state": "online", 00:13:23.299 "raid_level": "raid1", 00:13:23.299 "superblock": true, 00:13:23.299 "num_base_bdevs": 2, 00:13:23.299 "num_base_bdevs_discovered": 1, 00:13:23.299 "num_base_bdevs_operational": 1, 00:13:23.299 "base_bdevs_list": [ 00:13:23.299 { 00:13:23.299 "name": null, 00:13:23.299 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:23.299 "is_configured": false, 00:13:23.299 "data_offset": 0, 00:13:23.299 "data_size": 63488 00:13:23.299 }, 00:13:23.299 { 00:13:23.299 "name": "BaseBdev2", 00:13:23.299 "uuid": "84577df3-ff5e-5520-a176-05656fbbc918", 00:13:23.299 "is_configured": true, 00:13:23.299 "data_offset": 2048, 00:13:23.299 "data_size": 63488 00:13:23.299 } 00:13:23.299 ] 00:13:23.299 }' 00:13:23.299 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:23.559 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:23.559 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:23.559 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:23.559 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:13:23.559 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:23.559 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:23.559 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:23.559 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:13:23.559 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:23.559 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:23.559 [2024-11-19 15:19:13.730474] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:13:23.559 [2024-11-19 15:19:13.730562] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:23.559 [2024-11-19 15:19:13.730588] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:13:23.559 [2024-11-19 15:19:13.730598] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:23.559 [2024-11-19 15:19:13.731100] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:23.559 [2024-11-19 15:19:13.731125] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:13:23.559 [2024-11-19 15:19:13.731215] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:13:23.559 [2024-11-19 15:19:13.731236] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:13:23.559 [2024-11-19 15:19:13.731247] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:13:23.559 [2024-11-19 15:19:13.731271] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:13:23.559 BaseBdev1 00:13:23.559 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:23.559 15:19:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@775 -- # sleep 1 00:13:24.499 15:19:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:13:24.499 15:19:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:24.499 15:19:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:24.499 15:19:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:24.499 15:19:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:24.499 15:19:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:13:24.499 15:19:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:24.499 15:19:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:24.499 15:19:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:24.499 15:19:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:24.499 15:19:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:24.499 15:19:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:24.499 15:19:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:24.499 15:19:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:24.499 15:19:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:24.499 15:19:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:24.499 "name": "raid_bdev1", 00:13:24.499 "uuid": "92391f31-3636-4e85-85ea-1e816e5d07b9", 00:13:24.499 "strip_size_kb": 0, 00:13:24.499 "state": "online", 00:13:24.499 "raid_level": "raid1", 00:13:24.499 "superblock": true, 00:13:24.499 "num_base_bdevs": 2, 00:13:24.499 "num_base_bdevs_discovered": 1, 00:13:24.499 "num_base_bdevs_operational": 1, 00:13:24.499 "base_bdevs_list": [ 00:13:24.499 { 00:13:24.499 "name": null, 00:13:24.499 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:24.499 "is_configured": false, 00:13:24.499 "data_offset": 0, 00:13:24.499 "data_size": 63488 00:13:24.499 }, 00:13:24.499 { 00:13:24.499 "name": "BaseBdev2", 00:13:24.499 "uuid": "84577df3-ff5e-5520-a176-05656fbbc918", 00:13:24.499 "is_configured": true, 00:13:24.499 "data_offset": 2048, 00:13:24.499 "data_size": 63488 00:13:24.499 } 00:13:24.499 ] 00:13:24.499 }' 00:13:24.499 15:19:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:24.499 15:19:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:25.069 15:19:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:25.069 15:19:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:25.069 15:19:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:25.069 15:19:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:25.069 15:19:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:25.069 15:19:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:25.069 15:19:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:25.069 15:19:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:25.069 15:19:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:25.069 15:19:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:25.069 15:19:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:25.069 "name": "raid_bdev1", 00:13:25.069 "uuid": "92391f31-3636-4e85-85ea-1e816e5d07b9", 00:13:25.069 "strip_size_kb": 0, 00:13:25.069 "state": "online", 00:13:25.069 "raid_level": "raid1", 00:13:25.069 "superblock": true, 00:13:25.069 "num_base_bdevs": 2, 00:13:25.069 "num_base_bdevs_discovered": 1, 00:13:25.069 "num_base_bdevs_operational": 1, 00:13:25.069 "base_bdevs_list": [ 00:13:25.069 { 00:13:25.069 "name": null, 00:13:25.069 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:25.069 "is_configured": false, 00:13:25.069 "data_offset": 0, 00:13:25.069 "data_size": 63488 00:13:25.069 }, 00:13:25.069 { 00:13:25.069 "name": "BaseBdev2", 00:13:25.069 "uuid": "84577df3-ff5e-5520-a176-05656fbbc918", 00:13:25.069 "is_configured": true, 00:13:25.069 "data_offset": 2048, 00:13:25.069 "data_size": 63488 00:13:25.069 } 00:13:25.069 ] 00:13:25.069 }' 00:13:25.069 15:19:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:25.069 15:19:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:25.069 15:19:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:25.069 15:19:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:25.069 15:19:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:13:25.069 15:19:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@652 -- # local es=0 00:13:25.069 15:19:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:13:25.069 15:19:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:13:25.069 15:19:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:13:25.069 15:19:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:13:25.069 15:19:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:13:25.069 15:19:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:13:25.069 15:19:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:25.069 15:19:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:25.069 [2024-11-19 15:19:15.276137] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:25.069 [2024-11-19 15:19:15.276334] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:13:25.069 [2024-11-19 15:19:15.276375] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:13:25.069 request: 00:13:25.069 { 00:13:25.069 "base_bdev": "BaseBdev1", 00:13:25.069 "raid_bdev": "raid_bdev1", 00:13:25.069 "method": "bdev_raid_add_base_bdev", 00:13:25.069 "req_id": 1 00:13:25.069 } 00:13:25.069 Got JSON-RPC error response 00:13:25.069 response: 00:13:25.069 { 00:13:25.070 "code": -22, 00:13:25.070 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:13:25.070 } 00:13:25.070 15:19:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:13:25.070 15:19:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@655 -- # es=1 00:13:25.070 15:19:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:13:25.070 15:19:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:13:25.070 15:19:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:13:25.070 15:19:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@779 -- # sleep 1 00:13:26.009 15:19:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:13:26.009 15:19:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:26.009 15:19:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:26.009 15:19:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:26.009 15:19:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:26.009 15:19:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:13:26.009 15:19:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:26.009 15:19:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:26.009 15:19:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:26.009 15:19:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:26.009 15:19:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:26.009 15:19:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:26.009 15:19:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:26.009 15:19:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:26.009 15:19:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:26.009 15:19:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:26.009 "name": "raid_bdev1", 00:13:26.009 "uuid": "92391f31-3636-4e85-85ea-1e816e5d07b9", 00:13:26.009 "strip_size_kb": 0, 00:13:26.009 "state": "online", 00:13:26.009 "raid_level": "raid1", 00:13:26.009 "superblock": true, 00:13:26.009 "num_base_bdevs": 2, 00:13:26.009 "num_base_bdevs_discovered": 1, 00:13:26.009 "num_base_bdevs_operational": 1, 00:13:26.009 "base_bdevs_list": [ 00:13:26.009 { 00:13:26.009 "name": null, 00:13:26.009 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:26.009 "is_configured": false, 00:13:26.009 "data_offset": 0, 00:13:26.009 "data_size": 63488 00:13:26.009 }, 00:13:26.009 { 00:13:26.009 "name": "BaseBdev2", 00:13:26.009 "uuid": "84577df3-ff5e-5520-a176-05656fbbc918", 00:13:26.009 "is_configured": true, 00:13:26.009 "data_offset": 2048, 00:13:26.009 "data_size": 63488 00:13:26.009 } 00:13:26.009 ] 00:13:26.009 }' 00:13:26.009 15:19:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:26.009 15:19:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:26.579 15:19:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:26.579 15:19:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:26.579 15:19:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:26.579 15:19:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:26.579 15:19:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:26.579 15:19:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:26.579 15:19:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:26.579 15:19:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:26.579 15:19:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:26.579 15:19:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:26.579 15:19:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:26.579 "name": "raid_bdev1", 00:13:26.579 "uuid": "92391f31-3636-4e85-85ea-1e816e5d07b9", 00:13:26.579 "strip_size_kb": 0, 00:13:26.579 "state": "online", 00:13:26.579 "raid_level": "raid1", 00:13:26.579 "superblock": true, 00:13:26.579 "num_base_bdevs": 2, 00:13:26.579 "num_base_bdevs_discovered": 1, 00:13:26.579 "num_base_bdevs_operational": 1, 00:13:26.579 "base_bdevs_list": [ 00:13:26.579 { 00:13:26.579 "name": null, 00:13:26.579 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:26.579 "is_configured": false, 00:13:26.579 "data_offset": 0, 00:13:26.579 "data_size": 63488 00:13:26.579 }, 00:13:26.579 { 00:13:26.579 "name": "BaseBdev2", 00:13:26.579 "uuid": "84577df3-ff5e-5520-a176-05656fbbc918", 00:13:26.579 "is_configured": true, 00:13:26.579 "data_offset": 2048, 00:13:26.579 "data_size": 63488 00:13:26.579 } 00:13:26.579 ] 00:13:26.579 }' 00:13:26.579 15:19:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:26.579 15:19:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:26.579 15:19:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:26.579 15:19:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:26.579 15:19:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@784 -- # killprocess 87534 00:13:26.579 15:19:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@954 -- # '[' -z 87534 ']' 00:13:26.579 15:19:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@958 -- # kill -0 87534 00:13:26.579 15:19:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@959 -- # uname 00:13:26.579 15:19:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:13:26.579 15:19:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 87534 00:13:26.579 15:19:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:13:26.579 15:19:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:13:26.579 killing process with pid 87534 00:13:26.579 15:19:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@972 -- # echo 'killing process with pid 87534' 00:13:26.579 15:19:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@973 -- # kill 87534 00:13:26.579 Received shutdown signal, test time was about 16.495745 seconds 00:13:26.579 00:13:26.579 Latency(us) 00:13:26.579 [2024-11-19T15:19:16.918Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:26.579 [2024-11-19T15:19:16.918Z] =================================================================================================================== 00:13:26.579 [2024-11-19T15:19:16.918Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:13:26.579 [2024-11-19 15:19:16.859688] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:26.579 [2024-11-19 15:19:16.859844] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:26.579 15:19:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@978 -- # wait 87534 00:13:26.579 [2024-11-19 15:19:16.859914] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:26.579 [2024-11-19 15:19:16.859932] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:13:26.579 [2024-11-19 15:19:16.886472] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:26.839 15:19:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@786 -- # return 0 00:13:26.839 00:13:26.839 real 0m18.323s 00:13:26.839 user 0m24.384s 00:13:26.839 sys 0m2.044s 00:13:26.839 15:19:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1130 -- # xtrace_disable 00:13:26.839 15:19:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:26.839 ************************************ 00:13:26.839 END TEST raid_rebuild_test_sb_io 00:13:26.839 ************************************ 00:13:26.839 15:19:17 bdev_raid -- bdev/bdev_raid.sh@977 -- # for n in 2 4 00:13:26.839 15:19:17 bdev_raid -- bdev/bdev_raid.sh@978 -- # run_test raid_rebuild_test raid_rebuild_test raid1 4 false false true 00:13:26.839 15:19:17 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:13:26.839 15:19:17 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:13:26.839 15:19:17 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:26.839 ************************************ 00:13:26.839 START TEST raid_rebuild_test 00:13:26.839 ************************************ 00:13:26.839 15:19:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 4 false false true 00:13:26.839 15:19:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:13:26.839 15:19:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:13:26.839 15:19:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:13:26.839 15:19:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:13:26.839 15:19:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:13:26.839 15:19:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:13:26.839 15:19:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:26.839 15:19:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:13:26.839 15:19:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:26.839 15:19:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:26.839 15:19:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:13:26.839 15:19:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:26.839 15:19:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:26.839 15:19:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:13:26.839 15:19:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:26.839 15:19:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:26.839 15:19:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:13:26.839 15:19:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:26.839 15:19:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:26.839 15:19:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:13:26.839 15:19:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:13:27.099 15:19:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:13:27.099 15:19:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:13:27.099 15:19:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:13:27.099 15:19:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:13:27.099 15:19:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:13:27.099 15:19:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:13:27.099 15:19:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:13:27.099 15:19:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:13:27.099 15:19:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=88202 00:13:27.099 15:19:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:13:27.099 15:19:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 88202 00:13:27.099 15:19:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@835 -- # '[' -z 88202 ']' 00:13:27.099 15:19:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:27.099 15:19:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:13:27.099 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:27.099 15:19:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:27.099 15:19:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:13:27.099 15:19:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:27.099 I/O size of 3145728 is greater than zero copy threshold (65536). 00:13:27.099 Zero copy mechanism will not be used. 00:13:27.099 [2024-11-19 15:19:17.258995] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:13:27.099 [2024-11-19 15:19:17.259142] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid88202 ] 00:13:27.099 [2024-11-19 15:19:17.416214] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:27.359 [2024-11-19 15:19:17.442667] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:13:27.359 [2024-11-19 15:19:17.486411] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:27.359 [2024-11-19 15:19:17.486459] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:27.929 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:13:27.929 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@868 -- # return 0 00:13:27.929 15:19:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:27.929 15:19:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:13:27.929 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.929 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:27.929 BaseBdev1_malloc 00:13:27.929 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.929 15:19:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:13:27.929 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.929 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:27.929 [2024-11-19 15:19:18.089703] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:13:27.929 [2024-11-19 15:19:18.089787] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:27.929 [2024-11-19 15:19:18.089814] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:13:27.929 [2024-11-19 15:19:18.089836] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:27.929 [2024-11-19 15:19:18.092007] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:27.929 [2024-11-19 15:19:18.092045] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:13:27.929 BaseBdev1 00:13:27.929 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.929 15:19:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:27.929 15:19:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:13:27.929 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.929 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:27.929 BaseBdev2_malloc 00:13:27.929 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.929 15:19:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:13:27.929 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.929 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:27.930 [2024-11-19 15:19:18.118430] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:13:27.930 [2024-11-19 15:19:18.118500] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:27.930 [2024-11-19 15:19:18.118523] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:13:27.930 [2024-11-19 15:19:18.118532] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:27.930 [2024-11-19 15:19:18.120644] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:27.930 [2024-11-19 15:19:18.120694] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:13:27.930 BaseBdev2 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:27.930 BaseBdev3_malloc 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:27.930 [2024-11-19 15:19:18.147144] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:13:27.930 [2024-11-19 15:19:18.147220] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:27.930 [2024-11-19 15:19:18.147244] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:13:27.930 [2024-11-19 15:19:18.147254] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:27.930 [2024-11-19 15:19:18.149375] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:27.930 [2024-11-19 15:19:18.149412] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:13:27.930 BaseBdev3 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:27.930 BaseBdev4_malloc 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:27.930 [2024-11-19 15:19:18.190861] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:13:27.930 [2024-11-19 15:19:18.190941] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:27.930 [2024-11-19 15:19:18.190995] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:13:27.930 [2024-11-19 15:19:18.191014] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:27.930 [2024-11-19 15:19:18.194321] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:27.930 [2024-11-19 15:19:18.194380] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:13:27.930 BaseBdev4 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:27.930 spare_malloc 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:27.930 spare_delay 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:27.930 [2024-11-19 15:19:18.232121] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:27.930 [2024-11-19 15:19:18.232175] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:27.930 [2024-11-19 15:19:18.232212] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:13:27.930 [2024-11-19 15:19:18.232222] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:27.930 [2024-11-19 15:19:18.234323] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:27.930 [2024-11-19 15:19:18.234359] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:27.930 spare 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:27.930 [2024-11-19 15:19:18.244179] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:27.930 [2024-11-19 15:19:18.246083] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:27.930 [2024-11-19 15:19:18.246170] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:27.930 [2024-11-19 15:19:18.246221] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:13:27.930 [2024-11-19 15:19:18.246301] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:13:27.930 [2024-11-19 15:19:18.246317] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:13:27.930 [2024-11-19 15:19:18.246611] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:13:27.930 [2024-11-19 15:19:18.246759] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:13:27.930 [2024-11-19 15:19:18.246793] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:13:27.930 [2024-11-19 15:19:18.246944] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:27.930 15:19:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:28.191 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:28.191 15:19:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:28.191 "name": "raid_bdev1", 00:13:28.191 "uuid": "8ab4688e-282a-4ec8-a952-c1c94e6fd154", 00:13:28.191 "strip_size_kb": 0, 00:13:28.191 "state": "online", 00:13:28.191 "raid_level": "raid1", 00:13:28.191 "superblock": false, 00:13:28.191 "num_base_bdevs": 4, 00:13:28.191 "num_base_bdevs_discovered": 4, 00:13:28.191 "num_base_bdevs_operational": 4, 00:13:28.191 "base_bdevs_list": [ 00:13:28.191 { 00:13:28.191 "name": "BaseBdev1", 00:13:28.191 "uuid": "ba4b5080-a840-5373-aa53-af5b276a3321", 00:13:28.191 "is_configured": true, 00:13:28.191 "data_offset": 0, 00:13:28.191 "data_size": 65536 00:13:28.191 }, 00:13:28.191 { 00:13:28.191 "name": "BaseBdev2", 00:13:28.191 "uuid": "0b023a97-9d3d-5457-82a9-df81fef1624c", 00:13:28.191 "is_configured": true, 00:13:28.191 "data_offset": 0, 00:13:28.191 "data_size": 65536 00:13:28.191 }, 00:13:28.191 { 00:13:28.191 "name": "BaseBdev3", 00:13:28.191 "uuid": "2eff34af-a847-5ea3-a67b-cca348a1f241", 00:13:28.191 "is_configured": true, 00:13:28.191 "data_offset": 0, 00:13:28.191 "data_size": 65536 00:13:28.191 }, 00:13:28.191 { 00:13:28.191 "name": "BaseBdev4", 00:13:28.191 "uuid": "0703585a-c942-5cae-8021-342c206e20c2", 00:13:28.191 "is_configured": true, 00:13:28.191 "data_offset": 0, 00:13:28.191 "data_size": 65536 00:13:28.191 } 00:13:28.191 ] 00:13:28.191 }' 00:13:28.191 15:19:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:28.191 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:28.451 15:19:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:13:28.451 15:19:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:28.451 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:28.451 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:28.451 [2024-11-19 15:19:18.696165] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:28.451 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:28.451 15:19:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:13:28.451 15:19:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:13:28.451 15:19:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:28.451 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:28.451 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:28.451 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:28.451 15:19:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:13:28.451 15:19:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:13:28.451 15:19:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:13:28.451 15:19:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:13:28.451 15:19:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:13:28.451 15:19:18 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:28.451 15:19:18 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:13:28.451 15:19:18 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:28.451 15:19:18 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:13:28.451 15:19:18 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:28.451 15:19:18 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:13:28.451 15:19:18 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:28.451 15:19:18 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:28.451 15:19:18 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:13:28.711 [2024-11-19 15:19:18.927605] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:13:28.711 /dev/nbd0 00:13:28.711 15:19:18 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:13:28.711 15:19:18 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:13:28.711 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:13:28.711 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:13:28.711 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:28.711 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:28.711 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:13:28.711 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@877 -- # break 00:13:28.711 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:28.711 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:28.711 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:28.711 1+0 records in 00:13:28.711 1+0 records out 00:13:28.711 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000372182 s, 11.0 MB/s 00:13:28.711 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:28.711 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:13:28.711 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:28.711 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:28.711 15:19:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:13:28.711 15:19:18 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:28.711 15:19:18 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:28.711 15:19:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:13:28.711 15:19:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:13:28.711 15:19:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=65536 oflag=direct 00:13:35.288 65536+0 records in 00:13:35.288 65536+0 records out 00:13:35.288 33554432 bytes (34 MB, 32 MiB) copied, 5.365 s, 6.3 MB/s 00:13:35.288 15:19:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:13:35.288 15:19:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:35.288 15:19:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:13:35.288 15:19:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:35.288 15:19:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:13:35.288 15:19:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:35.288 15:19:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:13:35.288 [2024-11-19 15:19:24.538577] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:35.288 15:19:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:13:35.288 15:19:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:13:35.288 15:19:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:13:35.288 15:19:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:35.288 15:19:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:35.288 15:19:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:13:35.288 15:19:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:13:35.288 15:19:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:13:35.288 15:19:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:13:35.288 15:19:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:35.288 15:19:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.288 [2024-11-19 15:19:24.569797] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:35.288 15:19:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:35.288 15:19:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:13:35.288 15:19:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:35.288 15:19:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:35.288 15:19:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:35.288 15:19:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:35.288 15:19:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:35.288 15:19:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:35.288 15:19:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:35.288 15:19:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:35.288 15:19:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:35.288 15:19:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:35.289 15:19:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:35.289 15:19:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:35.289 15:19:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.289 15:19:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:35.289 15:19:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:35.289 "name": "raid_bdev1", 00:13:35.289 "uuid": "8ab4688e-282a-4ec8-a952-c1c94e6fd154", 00:13:35.289 "strip_size_kb": 0, 00:13:35.289 "state": "online", 00:13:35.289 "raid_level": "raid1", 00:13:35.289 "superblock": false, 00:13:35.289 "num_base_bdevs": 4, 00:13:35.289 "num_base_bdevs_discovered": 3, 00:13:35.289 "num_base_bdevs_operational": 3, 00:13:35.289 "base_bdevs_list": [ 00:13:35.289 { 00:13:35.289 "name": null, 00:13:35.289 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:35.289 "is_configured": false, 00:13:35.289 "data_offset": 0, 00:13:35.289 "data_size": 65536 00:13:35.289 }, 00:13:35.289 { 00:13:35.289 "name": "BaseBdev2", 00:13:35.289 "uuid": "0b023a97-9d3d-5457-82a9-df81fef1624c", 00:13:35.289 "is_configured": true, 00:13:35.289 "data_offset": 0, 00:13:35.289 "data_size": 65536 00:13:35.289 }, 00:13:35.289 { 00:13:35.289 "name": "BaseBdev3", 00:13:35.289 "uuid": "2eff34af-a847-5ea3-a67b-cca348a1f241", 00:13:35.289 "is_configured": true, 00:13:35.289 "data_offset": 0, 00:13:35.289 "data_size": 65536 00:13:35.289 }, 00:13:35.289 { 00:13:35.289 "name": "BaseBdev4", 00:13:35.289 "uuid": "0703585a-c942-5cae-8021-342c206e20c2", 00:13:35.289 "is_configured": true, 00:13:35.289 "data_offset": 0, 00:13:35.289 "data_size": 65536 00:13:35.289 } 00:13:35.289 ] 00:13:35.289 }' 00:13:35.289 15:19:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:35.289 15:19:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.289 15:19:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:35.289 15:19:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:35.289 15:19:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.289 [2024-11-19 15:19:25.049011] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:35.289 [2024-11-19 15:19:25.053508] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d063c0 00:13:35.289 15:19:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:35.289 15:19:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:13:35.289 [2024-11-19 15:19:25.055578] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:35.859 15:19:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:35.859 15:19:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:35.859 15:19:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:35.859 15:19:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:35.859 15:19:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:35.859 15:19:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:35.859 15:19:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:35.859 15:19:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:35.859 15:19:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.859 15:19:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:35.859 15:19:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:35.859 "name": "raid_bdev1", 00:13:35.859 "uuid": "8ab4688e-282a-4ec8-a952-c1c94e6fd154", 00:13:35.859 "strip_size_kb": 0, 00:13:35.859 "state": "online", 00:13:35.859 "raid_level": "raid1", 00:13:35.859 "superblock": false, 00:13:35.859 "num_base_bdevs": 4, 00:13:35.859 "num_base_bdevs_discovered": 4, 00:13:35.859 "num_base_bdevs_operational": 4, 00:13:35.859 "process": { 00:13:35.859 "type": "rebuild", 00:13:35.859 "target": "spare", 00:13:35.859 "progress": { 00:13:35.859 "blocks": 20480, 00:13:35.859 "percent": 31 00:13:35.859 } 00:13:35.859 }, 00:13:35.859 "base_bdevs_list": [ 00:13:35.859 { 00:13:35.859 "name": "spare", 00:13:35.859 "uuid": "b19625ef-4882-5cd1-87c4-2276662e3271", 00:13:35.859 "is_configured": true, 00:13:35.859 "data_offset": 0, 00:13:35.859 "data_size": 65536 00:13:35.859 }, 00:13:35.859 { 00:13:35.859 "name": "BaseBdev2", 00:13:35.859 "uuid": "0b023a97-9d3d-5457-82a9-df81fef1624c", 00:13:35.859 "is_configured": true, 00:13:35.859 "data_offset": 0, 00:13:35.859 "data_size": 65536 00:13:35.859 }, 00:13:35.859 { 00:13:35.859 "name": "BaseBdev3", 00:13:35.859 "uuid": "2eff34af-a847-5ea3-a67b-cca348a1f241", 00:13:35.859 "is_configured": true, 00:13:35.859 "data_offset": 0, 00:13:35.859 "data_size": 65536 00:13:35.859 }, 00:13:35.859 { 00:13:35.859 "name": "BaseBdev4", 00:13:35.859 "uuid": "0703585a-c942-5cae-8021-342c206e20c2", 00:13:35.859 "is_configured": true, 00:13:35.859 "data_offset": 0, 00:13:35.859 "data_size": 65536 00:13:35.859 } 00:13:35.859 ] 00:13:35.859 }' 00:13:35.859 15:19:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:35.859 15:19:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:35.859 15:19:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:35.859 15:19:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:35.859 15:19:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:13:35.859 15:19:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:35.859 15:19:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.859 [2024-11-19 15:19:26.188441] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:36.119 [2024-11-19 15:19:26.260321] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:36.119 [2024-11-19 15:19:26.260401] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:36.119 [2024-11-19 15:19:26.260423] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:36.119 [2024-11-19 15:19:26.260432] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:36.119 15:19:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:36.119 15:19:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:13:36.119 15:19:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:36.119 15:19:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:36.119 15:19:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:36.119 15:19:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:36.119 15:19:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:36.119 15:19:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:36.119 15:19:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:36.119 15:19:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:36.119 15:19:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:36.119 15:19:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:36.119 15:19:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:36.119 15:19:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:36.119 15:19:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:36.119 15:19:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:36.119 15:19:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:36.119 "name": "raid_bdev1", 00:13:36.119 "uuid": "8ab4688e-282a-4ec8-a952-c1c94e6fd154", 00:13:36.119 "strip_size_kb": 0, 00:13:36.119 "state": "online", 00:13:36.119 "raid_level": "raid1", 00:13:36.119 "superblock": false, 00:13:36.120 "num_base_bdevs": 4, 00:13:36.120 "num_base_bdevs_discovered": 3, 00:13:36.120 "num_base_bdevs_operational": 3, 00:13:36.120 "base_bdevs_list": [ 00:13:36.120 { 00:13:36.120 "name": null, 00:13:36.120 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:36.120 "is_configured": false, 00:13:36.120 "data_offset": 0, 00:13:36.120 "data_size": 65536 00:13:36.120 }, 00:13:36.120 { 00:13:36.120 "name": "BaseBdev2", 00:13:36.120 "uuid": "0b023a97-9d3d-5457-82a9-df81fef1624c", 00:13:36.120 "is_configured": true, 00:13:36.120 "data_offset": 0, 00:13:36.120 "data_size": 65536 00:13:36.120 }, 00:13:36.120 { 00:13:36.120 "name": "BaseBdev3", 00:13:36.120 "uuid": "2eff34af-a847-5ea3-a67b-cca348a1f241", 00:13:36.120 "is_configured": true, 00:13:36.120 "data_offset": 0, 00:13:36.120 "data_size": 65536 00:13:36.120 }, 00:13:36.120 { 00:13:36.120 "name": "BaseBdev4", 00:13:36.120 "uuid": "0703585a-c942-5cae-8021-342c206e20c2", 00:13:36.120 "is_configured": true, 00:13:36.120 "data_offset": 0, 00:13:36.120 "data_size": 65536 00:13:36.120 } 00:13:36.120 ] 00:13:36.120 }' 00:13:36.120 15:19:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:36.120 15:19:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:36.380 15:19:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:36.380 15:19:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:36.380 15:19:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:36.380 15:19:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:36.380 15:19:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:36.380 15:19:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:36.380 15:19:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:36.380 15:19:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:36.380 15:19:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:36.380 15:19:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:36.640 15:19:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:36.640 "name": "raid_bdev1", 00:13:36.640 "uuid": "8ab4688e-282a-4ec8-a952-c1c94e6fd154", 00:13:36.640 "strip_size_kb": 0, 00:13:36.640 "state": "online", 00:13:36.640 "raid_level": "raid1", 00:13:36.640 "superblock": false, 00:13:36.640 "num_base_bdevs": 4, 00:13:36.640 "num_base_bdevs_discovered": 3, 00:13:36.640 "num_base_bdevs_operational": 3, 00:13:36.640 "base_bdevs_list": [ 00:13:36.640 { 00:13:36.640 "name": null, 00:13:36.640 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:36.640 "is_configured": false, 00:13:36.640 "data_offset": 0, 00:13:36.640 "data_size": 65536 00:13:36.640 }, 00:13:36.640 { 00:13:36.640 "name": "BaseBdev2", 00:13:36.640 "uuid": "0b023a97-9d3d-5457-82a9-df81fef1624c", 00:13:36.640 "is_configured": true, 00:13:36.640 "data_offset": 0, 00:13:36.640 "data_size": 65536 00:13:36.640 }, 00:13:36.640 { 00:13:36.640 "name": "BaseBdev3", 00:13:36.640 "uuid": "2eff34af-a847-5ea3-a67b-cca348a1f241", 00:13:36.640 "is_configured": true, 00:13:36.640 "data_offset": 0, 00:13:36.640 "data_size": 65536 00:13:36.640 }, 00:13:36.640 { 00:13:36.640 "name": "BaseBdev4", 00:13:36.640 "uuid": "0703585a-c942-5cae-8021-342c206e20c2", 00:13:36.640 "is_configured": true, 00:13:36.640 "data_offset": 0, 00:13:36.640 "data_size": 65536 00:13:36.640 } 00:13:36.640 ] 00:13:36.640 }' 00:13:36.640 15:19:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:36.640 15:19:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:36.640 15:19:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:36.640 15:19:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:36.640 15:19:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:36.640 15:19:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:36.640 15:19:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:36.640 [2024-11-19 15:19:26.804075] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:36.640 [2024-11-19 15:19:26.807906] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d06490 00:13:36.640 [2024-11-19 15:19:26.809770] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:36.640 15:19:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:36.640 15:19:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:13:37.579 15:19:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:37.579 15:19:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:37.579 15:19:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:37.579 15:19:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:37.579 15:19:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:37.579 15:19:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:37.579 15:19:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:37.579 15:19:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:37.579 15:19:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:37.579 15:19:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:37.579 15:19:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:37.579 "name": "raid_bdev1", 00:13:37.579 "uuid": "8ab4688e-282a-4ec8-a952-c1c94e6fd154", 00:13:37.579 "strip_size_kb": 0, 00:13:37.579 "state": "online", 00:13:37.579 "raid_level": "raid1", 00:13:37.579 "superblock": false, 00:13:37.579 "num_base_bdevs": 4, 00:13:37.579 "num_base_bdevs_discovered": 4, 00:13:37.579 "num_base_bdevs_operational": 4, 00:13:37.579 "process": { 00:13:37.579 "type": "rebuild", 00:13:37.579 "target": "spare", 00:13:37.579 "progress": { 00:13:37.579 "blocks": 20480, 00:13:37.579 "percent": 31 00:13:37.579 } 00:13:37.579 }, 00:13:37.579 "base_bdevs_list": [ 00:13:37.579 { 00:13:37.579 "name": "spare", 00:13:37.579 "uuid": "b19625ef-4882-5cd1-87c4-2276662e3271", 00:13:37.579 "is_configured": true, 00:13:37.579 "data_offset": 0, 00:13:37.579 "data_size": 65536 00:13:37.579 }, 00:13:37.579 { 00:13:37.579 "name": "BaseBdev2", 00:13:37.579 "uuid": "0b023a97-9d3d-5457-82a9-df81fef1624c", 00:13:37.579 "is_configured": true, 00:13:37.579 "data_offset": 0, 00:13:37.579 "data_size": 65536 00:13:37.579 }, 00:13:37.579 { 00:13:37.579 "name": "BaseBdev3", 00:13:37.579 "uuid": "2eff34af-a847-5ea3-a67b-cca348a1f241", 00:13:37.579 "is_configured": true, 00:13:37.579 "data_offset": 0, 00:13:37.579 "data_size": 65536 00:13:37.579 }, 00:13:37.579 { 00:13:37.579 "name": "BaseBdev4", 00:13:37.579 "uuid": "0703585a-c942-5cae-8021-342c206e20c2", 00:13:37.579 "is_configured": true, 00:13:37.579 "data_offset": 0, 00:13:37.579 "data_size": 65536 00:13:37.579 } 00:13:37.579 ] 00:13:37.579 }' 00:13:37.579 15:19:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:37.579 15:19:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:37.579 15:19:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:37.840 15:19:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:37.840 15:19:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:13:37.840 15:19:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:13:37.840 15:19:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:13:37.840 15:19:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:13:37.840 15:19:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:13:37.840 15:19:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:37.840 15:19:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:37.840 [2024-11-19 15:19:27.944616] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:13:37.840 [2024-11-19 15:19:28.013909] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000d06490 00:13:37.840 15:19:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:37.840 15:19:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:13:37.840 15:19:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:13:37.840 15:19:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:37.840 15:19:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:37.840 15:19:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:37.840 15:19:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:37.840 15:19:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:37.840 15:19:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:37.840 15:19:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:37.840 15:19:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:37.840 15:19:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:37.840 15:19:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:37.840 15:19:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:37.840 "name": "raid_bdev1", 00:13:37.840 "uuid": "8ab4688e-282a-4ec8-a952-c1c94e6fd154", 00:13:37.840 "strip_size_kb": 0, 00:13:37.840 "state": "online", 00:13:37.840 "raid_level": "raid1", 00:13:37.840 "superblock": false, 00:13:37.840 "num_base_bdevs": 4, 00:13:37.840 "num_base_bdevs_discovered": 3, 00:13:37.840 "num_base_bdevs_operational": 3, 00:13:37.840 "process": { 00:13:37.840 "type": "rebuild", 00:13:37.840 "target": "spare", 00:13:37.840 "progress": { 00:13:37.840 "blocks": 24576, 00:13:37.840 "percent": 37 00:13:37.840 } 00:13:37.840 }, 00:13:37.840 "base_bdevs_list": [ 00:13:37.840 { 00:13:37.840 "name": "spare", 00:13:37.840 "uuid": "b19625ef-4882-5cd1-87c4-2276662e3271", 00:13:37.840 "is_configured": true, 00:13:37.840 "data_offset": 0, 00:13:37.840 "data_size": 65536 00:13:37.840 }, 00:13:37.840 { 00:13:37.840 "name": null, 00:13:37.840 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:37.840 "is_configured": false, 00:13:37.840 "data_offset": 0, 00:13:37.840 "data_size": 65536 00:13:37.840 }, 00:13:37.840 { 00:13:37.840 "name": "BaseBdev3", 00:13:37.840 "uuid": "2eff34af-a847-5ea3-a67b-cca348a1f241", 00:13:37.840 "is_configured": true, 00:13:37.840 "data_offset": 0, 00:13:37.840 "data_size": 65536 00:13:37.840 }, 00:13:37.840 { 00:13:37.840 "name": "BaseBdev4", 00:13:37.840 "uuid": "0703585a-c942-5cae-8021-342c206e20c2", 00:13:37.840 "is_configured": true, 00:13:37.840 "data_offset": 0, 00:13:37.840 "data_size": 65536 00:13:37.840 } 00:13:37.840 ] 00:13:37.840 }' 00:13:37.840 15:19:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:37.840 15:19:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:37.840 15:19:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:37.840 15:19:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:37.840 15:19:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=366 00:13:37.840 15:19:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:37.840 15:19:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:37.840 15:19:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:37.840 15:19:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:37.840 15:19:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:37.840 15:19:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:37.840 15:19:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:37.840 15:19:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:37.840 15:19:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:37.840 15:19:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:37.840 15:19:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:38.100 15:19:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:38.100 "name": "raid_bdev1", 00:13:38.100 "uuid": "8ab4688e-282a-4ec8-a952-c1c94e6fd154", 00:13:38.100 "strip_size_kb": 0, 00:13:38.100 "state": "online", 00:13:38.100 "raid_level": "raid1", 00:13:38.100 "superblock": false, 00:13:38.100 "num_base_bdevs": 4, 00:13:38.100 "num_base_bdevs_discovered": 3, 00:13:38.100 "num_base_bdevs_operational": 3, 00:13:38.100 "process": { 00:13:38.100 "type": "rebuild", 00:13:38.100 "target": "spare", 00:13:38.100 "progress": { 00:13:38.100 "blocks": 26624, 00:13:38.100 "percent": 40 00:13:38.100 } 00:13:38.100 }, 00:13:38.100 "base_bdevs_list": [ 00:13:38.100 { 00:13:38.100 "name": "spare", 00:13:38.100 "uuid": "b19625ef-4882-5cd1-87c4-2276662e3271", 00:13:38.100 "is_configured": true, 00:13:38.100 "data_offset": 0, 00:13:38.100 "data_size": 65536 00:13:38.100 }, 00:13:38.100 { 00:13:38.100 "name": null, 00:13:38.100 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:38.100 "is_configured": false, 00:13:38.100 "data_offset": 0, 00:13:38.100 "data_size": 65536 00:13:38.100 }, 00:13:38.100 { 00:13:38.100 "name": "BaseBdev3", 00:13:38.100 "uuid": "2eff34af-a847-5ea3-a67b-cca348a1f241", 00:13:38.100 "is_configured": true, 00:13:38.100 "data_offset": 0, 00:13:38.100 "data_size": 65536 00:13:38.100 }, 00:13:38.100 { 00:13:38.100 "name": "BaseBdev4", 00:13:38.100 "uuid": "0703585a-c942-5cae-8021-342c206e20c2", 00:13:38.100 "is_configured": true, 00:13:38.100 "data_offset": 0, 00:13:38.100 "data_size": 65536 00:13:38.100 } 00:13:38.100 ] 00:13:38.100 }' 00:13:38.100 15:19:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:38.100 15:19:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:38.101 15:19:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:38.101 15:19:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:38.101 15:19:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:39.040 15:19:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:39.040 15:19:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:39.040 15:19:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:39.040 15:19:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:39.040 15:19:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:39.040 15:19:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:39.040 15:19:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:39.040 15:19:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:39.040 15:19:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:39.040 15:19:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:39.040 15:19:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:39.040 15:19:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:39.040 "name": "raid_bdev1", 00:13:39.040 "uuid": "8ab4688e-282a-4ec8-a952-c1c94e6fd154", 00:13:39.040 "strip_size_kb": 0, 00:13:39.040 "state": "online", 00:13:39.040 "raid_level": "raid1", 00:13:39.040 "superblock": false, 00:13:39.040 "num_base_bdevs": 4, 00:13:39.040 "num_base_bdevs_discovered": 3, 00:13:39.040 "num_base_bdevs_operational": 3, 00:13:39.040 "process": { 00:13:39.040 "type": "rebuild", 00:13:39.040 "target": "spare", 00:13:39.040 "progress": { 00:13:39.040 "blocks": 49152, 00:13:39.040 "percent": 75 00:13:39.040 } 00:13:39.040 }, 00:13:39.041 "base_bdevs_list": [ 00:13:39.041 { 00:13:39.041 "name": "spare", 00:13:39.041 "uuid": "b19625ef-4882-5cd1-87c4-2276662e3271", 00:13:39.041 "is_configured": true, 00:13:39.041 "data_offset": 0, 00:13:39.041 "data_size": 65536 00:13:39.041 }, 00:13:39.041 { 00:13:39.041 "name": null, 00:13:39.041 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:39.041 "is_configured": false, 00:13:39.041 "data_offset": 0, 00:13:39.041 "data_size": 65536 00:13:39.041 }, 00:13:39.041 { 00:13:39.041 "name": "BaseBdev3", 00:13:39.041 "uuid": "2eff34af-a847-5ea3-a67b-cca348a1f241", 00:13:39.041 "is_configured": true, 00:13:39.041 "data_offset": 0, 00:13:39.041 "data_size": 65536 00:13:39.041 }, 00:13:39.041 { 00:13:39.041 "name": "BaseBdev4", 00:13:39.041 "uuid": "0703585a-c942-5cae-8021-342c206e20c2", 00:13:39.041 "is_configured": true, 00:13:39.041 "data_offset": 0, 00:13:39.041 "data_size": 65536 00:13:39.041 } 00:13:39.041 ] 00:13:39.041 }' 00:13:39.041 15:19:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:39.301 15:19:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:39.301 15:19:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:39.301 15:19:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:39.301 15:19:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:39.870 [2024-11-19 15:19:30.021375] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:13:39.870 [2024-11-19 15:19:30.021490] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:13:39.870 [2024-11-19 15:19:30.021541] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:40.130 15:19:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:40.130 15:19:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:40.130 15:19:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:40.130 15:19:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:40.130 15:19:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:40.130 15:19:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:40.130 15:19:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:40.130 15:19:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:40.130 15:19:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:40.130 15:19:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:40.390 15:19:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:40.390 15:19:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:40.390 "name": "raid_bdev1", 00:13:40.390 "uuid": "8ab4688e-282a-4ec8-a952-c1c94e6fd154", 00:13:40.390 "strip_size_kb": 0, 00:13:40.390 "state": "online", 00:13:40.390 "raid_level": "raid1", 00:13:40.390 "superblock": false, 00:13:40.390 "num_base_bdevs": 4, 00:13:40.390 "num_base_bdevs_discovered": 3, 00:13:40.390 "num_base_bdevs_operational": 3, 00:13:40.390 "base_bdevs_list": [ 00:13:40.390 { 00:13:40.390 "name": "spare", 00:13:40.390 "uuid": "b19625ef-4882-5cd1-87c4-2276662e3271", 00:13:40.390 "is_configured": true, 00:13:40.390 "data_offset": 0, 00:13:40.390 "data_size": 65536 00:13:40.390 }, 00:13:40.390 { 00:13:40.390 "name": null, 00:13:40.390 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:40.390 "is_configured": false, 00:13:40.390 "data_offset": 0, 00:13:40.390 "data_size": 65536 00:13:40.390 }, 00:13:40.390 { 00:13:40.390 "name": "BaseBdev3", 00:13:40.390 "uuid": "2eff34af-a847-5ea3-a67b-cca348a1f241", 00:13:40.390 "is_configured": true, 00:13:40.390 "data_offset": 0, 00:13:40.390 "data_size": 65536 00:13:40.390 }, 00:13:40.390 { 00:13:40.390 "name": "BaseBdev4", 00:13:40.390 "uuid": "0703585a-c942-5cae-8021-342c206e20c2", 00:13:40.390 "is_configured": true, 00:13:40.390 "data_offset": 0, 00:13:40.390 "data_size": 65536 00:13:40.390 } 00:13:40.390 ] 00:13:40.390 }' 00:13:40.390 15:19:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:40.390 15:19:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:13:40.390 15:19:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:40.390 15:19:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:13:40.390 15:19:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:13:40.390 15:19:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:40.390 15:19:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:40.390 15:19:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:40.390 15:19:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:40.390 15:19:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:40.390 15:19:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:40.390 15:19:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:40.390 15:19:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:40.390 15:19:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:40.390 15:19:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:40.390 15:19:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:40.390 "name": "raid_bdev1", 00:13:40.390 "uuid": "8ab4688e-282a-4ec8-a952-c1c94e6fd154", 00:13:40.390 "strip_size_kb": 0, 00:13:40.390 "state": "online", 00:13:40.390 "raid_level": "raid1", 00:13:40.390 "superblock": false, 00:13:40.390 "num_base_bdevs": 4, 00:13:40.390 "num_base_bdevs_discovered": 3, 00:13:40.390 "num_base_bdevs_operational": 3, 00:13:40.390 "base_bdevs_list": [ 00:13:40.390 { 00:13:40.390 "name": "spare", 00:13:40.390 "uuid": "b19625ef-4882-5cd1-87c4-2276662e3271", 00:13:40.390 "is_configured": true, 00:13:40.390 "data_offset": 0, 00:13:40.390 "data_size": 65536 00:13:40.390 }, 00:13:40.390 { 00:13:40.390 "name": null, 00:13:40.390 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:40.390 "is_configured": false, 00:13:40.390 "data_offset": 0, 00:13:40.390 "data_size": 65536 00:13:40.390 }, 00:13:40.390 { 00:13:40.390 "name": "BaseBdev3", 00:13:40.390 "uuid": "2eff34af-a847-5ea3-a67b-cca348a1f241", 00:13:40.390 "is_configured": true, 00:13:40.390 "data_offset": 0, 00:13:40.390 "data_size": 65536 00:13:40.390 }, 00:13:40.390 { 00:13:40.390 "name": "BaseBdev4", 00:13:40.390 "uuid": "0703585a-c942-5cae-8021-342c206e20c2", 00:13:40.390 "is_configured": true, 00:13:40.390 "data_offset": 0, 00:13:40.390 "data_size": 65536 00:13:40.390 } 00:13:40.390 ] 00:13:40.390 }' 00:13:40.390 15:19:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:40.390 15:19:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:40.390 15:19:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:40.650 15:19:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:40.650 15:19:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:13:40.650 15:19:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:40.650 15:19:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:40.650 15:19:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:40.650 15:19:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:40.650 15:19:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:40.650 15:19:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:40.650 15:19:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:40.650 15:19:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:40.650 15:19:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:40.650 15:19:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:40.650 15:19:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:40.650 15:19:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:40.650 15:19:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:40.650 15:19:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:40.650 15:19:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:40.650 "name": "raid_bdev1", 00:13:40.650 "uuid": "8ab4688e-282a-4ec8-a952-c1c94e6fd154", 00:13:40.650 "strip_size_kb": 0, 00:13:40.650 "state": "online", 00:13:40.650 "raid_level": "raid1", 00:13:40.650 "superblock": false, 00:13:40.650 "num_base_bdevs": 4, 00:13:40.650 "num_base_bdevs_discovered": 3, 00:13:40.650 "num_base_bdevs_operational": 3, 00:13:40.650 "base_bdevs_list": [ 00:13:40.650 { 00:13:40.650 "name": "spare", 00:13:40.650 "uuid": "b19625ef-4882-5cd1-87c4-2276662e3271", 00:13:40.650 "is_configured": true, 00:13:40.650 "data_offset": 0, 00:13:40.650 "data_size": 65536 00:13:40.650 }, 00:13:40.650 { 00:13:40.650 "name": null, 00:13:40.650 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:40.650 "is_configured": false, 00:13:40.650 "data_offset": 0, 00:13:40.650 "data_size": 65536 00:13:40.650 }, 00:13:40.650 { 00:13:40.650 "name": "BaseBdev3", 00:13:40.650 "uuid": "2eff34af-a847-5ea3-a67b-cca348a1f241", 00:13:40.650 "is_configured": true, 00:13:40.650 "data_offset": 0, 00:13:40.650 "data_size": 65536 00:13:40.650 }, 00:13:40.650 { 00:13:40.650 "name": "BaseBdev4", 00:13:40.650 "uuid": "0703585a-c942-5cae-8021-342c206e20c2", 00:13:40.650 "is_configured": true, 00:13:40.650 "data_offset": 0, 00:13:40.650 "data_size": 65536 00:13:40.650 } 00:13:40.650 ] 00:13:40.650 }' 00:13:40.650 15:19:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:40.650 15:19:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:40.910 15:19:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:13:40.910 15:19:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:40.910 15:19:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:40.910 [2024-11-19 15:19:31.103748] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:40.910 [2024-11-19 15:19:31.103783] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:40.910 [2024-11-19 15:19:31.103903] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:40.911 [2024-11-19 15:19:31.104009] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:40.911 [2024-11-19 15:19:31.104033] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:13:40.911 15:19:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:40.911 15:19:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:40.911 15:19:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:40.911 15:19:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:13:40.911 15:19:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:40.911 15:19:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:40.911 15:19:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:13:40.911 15:19:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:13:40.911 15:19:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:13:40.911 15:19:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:13:40.911 15:19:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:40.911 15:19:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:13:40.911 15:19:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:40.911 15:19:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:13:40.911 15:19:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:40.911 15:19:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:13:40.911 15:19:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:40.911 15:19:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:13:40.911 15:19:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:13:41.171 /dev/nbd0 00:13:41.171 15:19:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:13:41.171 15:19:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:13:41.171 15:19:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:13:41.171 15:19:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:13:41.171 15:19:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:41.171 15:19:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:41.171 15:19:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:13:41.171 15:19:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@877 -- # break 00:13:41.171 15:19:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:41.171 15:19:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:41.171 15:19:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:41.171 1+0 records in 00:13:41.171 1+0 records out 00:13:41.171 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000404216 s, 10.1 MB/s 00:13:41.171 15:19:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:41.171 15:19:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:13:41.171 15:19:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:41.171 15:19:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:41.171 15:19:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:13:41.171 15:19:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:41.171 15:19:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:13:41.171 15:19:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:13:41.432 /dev/nbd1 00:13:41.432 15:19:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:13:41.432 15:19:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:13:41.432 15:19:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:13:41.432 15:19:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:13:41.432 15:19:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:41.432 15:19:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:41.432 15:19:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:13:41.432 15:19:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@877 -- # break 00:13:41.432 15:19:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:41.432 15:19:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:41.432 15:19:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:41.432 1+0 records in 00:13:41.432 1+0 records out 00:13:41.432 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000252867 s, 16.2 MB/s 00:13:41.432 15:19:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:41.432 15:19:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:13:41.432 15:19:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:41.432 15:19:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:41.432 15:19:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:13:41.432 15:19:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:41.432 15:19:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:13:41.432 15:19:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:13:41.432 15:19:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:13:41.432 15:19:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:41.432 15:19:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:13:41.432 15:19:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:41.432 15:19:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:13:41.432 15:19:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:41.432 15:19:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:13:41.692 15:19:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:13:41.692 15:19:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:13:41.692 15:19:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:13:41.692 15:19:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:41.692 15:19:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:41.692 15:19:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:13:41.692 15:19:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:13:41.692 15:19:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:13:41.692 15:19:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:41.692 15:19:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:13:41.952 15:19:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:13:41.952 15:19:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:13:41.952 15:19:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:13:41.952 15:19:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:41.952 15:19:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:41.952 15:19:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:13:41.952 15:19:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:13:41.952 15:19:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:13:41.952 15:19:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:13:41.952 15:19:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 88202 00:13:41.952 15:19:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@954 -- # '[' -z 88202 ']' 00:13:41.952 15:19:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@958 -- # kill -0 88202 00:13:41.952 15:19:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@959 -- # uname 00:13:41.952 15:19:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:13:41.952 15:19:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 88202 00:13:41.952 15:19:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:13:41.952 15:19:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:13:41.952 killing process with pid 88202 00:13:41.952 15:19:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 88202' 00:13:41.952 15:19:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@973 -- # kill 88202 00:13:41.952 Received shutdown signal, test time was about 60.000000 seconds 00:13:41.952 00:13:41.952 Latency(us) 00:13:41.952 [2024-11-19T15:19:32.291Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:41.952 [2024-11-19T15:19:32.291Z] =================================================================================================================== 00:13:41.952 [2024-11-19T15:19:32.291Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:13:41.952 [2024-11-19 15:19:32.155776] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:41.952 15:19:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@978 -- # wait 88202 00:13:41.952 [2024-11-19 15:19:32.208303] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:42.213 15:19:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:13:42.213 00:13:42.213 real 0m15.248s 00:13:42.213 user 0m17.349s 00:13:42.213 sys 0m2.952s 00:13:42.213 15:19:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:13:42.213 15:19:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:42.213 ************************************ 00:13:42.213 END TEST raid_rebuild_test 00:13:42.213 ************************************ 00:13:42.213 15:19:32 bdev_raid -- bdev/bdev_raid.sh@979 -- # run_test raid_rebuild_test_sb raid_rebuild_test raid1 4 true false true 00:13:42.213 15:19:32 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:13:42.213 15:19:32 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:13:42.213 15:19:32 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:42.213 ************************************ 00:13:42.213 START TEST raid_rebuild_test_sb 00:13:42.213 ************************************ 00:13:42.213 15:19:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 4 true false true 00:13:42.213 15:19:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:13:42.213 15:19:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:13:42.213 15:19:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:13:42.213 15:19:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:13:42.213 15:19:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:13:42.213 15:19:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:13:42.213 15:19:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:42.213 15:19:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:13:42.213 15:19:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:42.213 15:19:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:42.213 15:19:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:13:42.213 15:19:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:42.213 15:19:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:42.213 15:19:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:13:42.213 15:19:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:42.213 15:19:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:42.213 15:19:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:13:42.213 15:19:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:42.213 15:19:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:42.213 15:19:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:13:42.213 15:19:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:13:42.213 15:19:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:13:42.213 15:19:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:13:42.213 15:19:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:13:42.213 15:19:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:13:42.213 15:19:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:13:42.213 15:19:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:13:42.213 15:19:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:13:42.213 15:19:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:13:42.213 15:19:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:13:42.213 15:19:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=88627 00:13:42.213 15:19:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:13:42.213 15:19:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 88627 00:13:42.213 15:19:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@835 -- # '[' -z 88627 ']' 00:13:42.213 15:19:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:42.213 15:19:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:13:42.213 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:42.213 15:19:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:42.213 15:19:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:13:42.213 15:19:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:42.473 I/O size of 3145728 is greater than zero copy threshold (65536). 00:13:42.473 Zero copy mechanism will not be used. 00:13:42.473 [2024-11-19 15:19:32.590357] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:13:42.473 [2024-11-19 15:19:32.590500] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid88627 ] 00:13:42.473 [2024-11-19 15:19:32.745164] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:42.473 [2024-11-19 15:19:32.772204] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:13:42.735 [2024-11-19 15:19:32.816158] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:42.735 [2024-11-19 15:19:32.816206] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@868 -- # return 0 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:43.338 BaseBdev1_malloc 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:43.338 [2024-11-19 15:19:33.423641] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:13:43.338 [2024-11-19 15:19:33.423743] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:43.338 [2024-11-19 15:19:33.423774] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:13:43.338 [2024-11-19 15:19:33.423788] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:43.338 [2024-11-19 15:19:33.425885] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:43.338 [2024-11-19 15:19:33.425924] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:13:43.338 BaseBdev1 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:43.338 BaseBdev2_malloc 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:43.338 [2024-11-19 15:19:33.452501] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:13:43.338 [2024-11-19 15:19:33.452559] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:43.338 [2024-11-19 15:19:33.452598] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:13:43.338 [2024-11-19 15:19:33.452608] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:43.338 [2024-11-19 15:19:33.454689] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:43.338 [2024-11-19 15:19:33.454732] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:13:43.338 BaseBdev2 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:43.338 BaseBdev3_malloc 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:43.338 [2024-11-19 15:19:33.481441] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:13:43.338 [2024-11-19 15:19:33.481515] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:43.338 [2024-11-19 15:19:33.481539] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:13:43.338 [2024-11-19 15:19:33.481550] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:43.338 [2024-11-19 15:19:33.483644] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:43.338 [2024-11-19 15:19:33.483680] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:13:43.338 BaseBdev3 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:43.338 BaseBdev4_malloc 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:43.338 [2024-11-19 15:19:33.526994] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:13:43.338 [2024-11-19 15:19:33.527080] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:43.338 [2024-11-19 15:19:33.527142] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:13:43.338 [2024-11-19 15:19:33.527163] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:43.338 [2024-11-19 15:19:33.531000] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:43.338 [2024-11-19 15:19:33.531063] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:13:43.338 BaseBdev4 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:43.338 spare_malloc 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:43.338 spare_delay 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:43.338 [2024-11-19 15:19:33.568663] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:43.338 [2024-11-19 15:19:33.568718] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:43.338 [2024-11-19 15:19:33.568755] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:13:43.338 [2024-11-19 15:19:33.568765] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:43.338 [2024-11-19 15:19:33.570887] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:43.338 [2024-11-19 15:19:33.570924] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:43.338 spare 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.338 15:19:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:43.338 [2024-11-19 15:19:33.580719] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:43.338 [2024-11-19 15:19:33.582616] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:43.338 [2024-11-19 15:19:33.582706] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:43.339 [2024-11-19 15:19:33.582758] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:13:43.339 [2024-11-19 15:19:33.582956] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:13:43.339 [2024-11-19 15:19:33.582994] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:13:43.339 [2024-11-19 15:19:33.583245] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:13:43.339 [2024-11-19 15:19:33.583405] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:13:43.339 [2024-11-19 15:19:33.583433] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:13:43.339 [2024-11-19 15:19:33.583577] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:43.339 15:19:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.339 15:19:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:13:43.339 15:19:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:43.339 15:19:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:43.339 15:19:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:43.339 15:19:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:43.339 15:19:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:13:43.339 15:19:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:43.339 15:19:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:43.339 15:19:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:43.339 15:19:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:43.339 15:19:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:43.339 15:19:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:43.339 15:19:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.339 15:19:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:43.339 15:19:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.339 15:19:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:43.339 "name": "raid_bdev1", 00:13:43.339 "uuid": "499db5d6-2ee5-45f5-b8aa-305d1e2ff852", 00:13:43.339 "strip_size_kb": 0, 00:13:43.339 "state": "online", 00:13:43.339 "raid_level": "raid1", 00:13:43.339 "superblock": true, 00:13:43.339 "num_base_bdevs": 4, 00:13:43.339 "num_base_bdevs_discovered": 4, 00:13:43.339 "num_base_bdevs_operational": 4, 00:13:43.339 "base_bdevs_list": [ 00:13:43.339 { 00:13:43.339 "name": "BaseBdev1", 00:13:43.339 "uuid": "13b30770-a237-5088-8d19-ca655c8dcc71", 00:13:43.339 "is_configured": true, 00:13:43.339 "data_offset": 2048, 00:13:43.339 "data_size": 63488 00:13:43.339 }, 00:13:43.339 { 00:13:43.339 "name": "BaseBdev2", 00:13:43.339 "uuid": "a468af4b-33fd-5634-99dc-44fc225c050f", 00:13:43.339 "is_configured": true, 00:13:43.339 "data_offset": 2048, 00:13:43.339 "data_size": 63488 00:13:43.339 }, 00:13:43.339 { 00:13:43.339 "name": "BaseBdev3", 00:13:43.339 "uuid": "bba59b1b-a5ac-5586-a17a-fa7c79f7e077", 00:13:43.339 "is_configured": true, 00:13:43.339 "data_offset": 2048, 00:13:43.339 "data_size": 63488 00:13:43.339 }, 00:13:43.339 { 00:13:43.339 "name": "BaseBdev4", 00:13:43.339 "uuid": "bcf97a61-829b-56dd-ae06-5f30dda1bd20", 00:13:43.339 "is_configured": true, 00:13:43.339 "data_offset": 2048, 00:13:43.339 "data_size": 63488 00:13:43.339 } 00:13:43.339 ] 00:13:43.339 }' 00:13:43.339 15:19:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:43.339 15:19:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:43.924 15:19:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:13:43.924 15:19:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:43.924 15:19:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.924 15:19:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:43.924 [2024-11-19 15:19:34.000344] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:43.924 15:19:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.924 15:19:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:13:43.924 15:19:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:43.924 15:19:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.924 15:19:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:43.924 15:19:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:13:43.924 15:19:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.924 15:19:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:13:43.924 15:19:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:13:43.924 15:19:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:13:43.924 15:19:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:13:43.924 15:19:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:13:43.924 15:19:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:43.924 15:19:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:13:43.924 15:19:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:43.924 15:19:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:13:43.924 15:19:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:43.924 15:19:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:13:43.924 15:19:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:43.924 15:19:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:43.924 15:19:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:13:43.924 [2024-11-19 15:19:34.255622] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:13:44.184 /dev/nbd0 00:13:44.184 15:19:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:13:44.184 15:19:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:13:44.184 15:19:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:13:44.184 15:19:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:13:44.184 15:19:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:44.184 15:19:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:44.184 15:19:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:13:44.184 15:19:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:13:44.184 15:19:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:44.184 15:19:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:44.184 15:19:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:44.184 1+0 records in 00:13:44.184 1+0 records out 00:13:44.184 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000371257 s, 11.0 MB/s 00:13:44.184 15:19:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:44.184 15:19:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:13:44.184 15:19:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:44.184 15:19:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:44.184 15:19:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:13:44.185 15:19:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:44.185 15:19:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:44.185 15:19:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:13:44.185 15:19:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:13:44.185 15:19:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=63488 oflag=direct 00:13:50.759 63488+0 records in 00:13:50.759 63488+0 records out 00:13:50.759 32505856 bytes (33 MB, 31 MiB) copied, 5.74685 s, 5.7 MB/s 00:13:50.759 15:19:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:13:50.759 15:19:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:50.759 15:19:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:13:50.759 15:19:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:50.759 15:19:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:13:50.759 15:19:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:50.759 15:19:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:13:50.759 15:19:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:13:50.759 [2024-11-19 15:19:40.271613] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:50.759 15:19:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:13:50.759 15:19:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:13:50.759 15:19:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:50.759 15:19:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:50.759 15:19:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:13:50.759 15:19:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:13:50.759 15:19:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:13:50.759 15:19:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:13:50.759 15:19:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.759 15:19:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:50.759 [2024-11-19 15:19:40.288402] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:50.759 15:19:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.759 15:19:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:13:50.759 15:19:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:50.759 15:19:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:50.759 15:19:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:50.759 15:19:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:50.759 15:19:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:50.759 15:19:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:50.759 15:19:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:50.759 15:19:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:50.759 15:19:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:50.759 15:19:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:50.760 15:19:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.760 15:19:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:50.760 15:19:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:50.760 15:19:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.760 15:19:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:50.760 "name": "raid_bdev1", 00:13:50.760 "uuid": "499db5d6-2ee5-45f5-b8aa-305d1e2ff852", 00:13:50.760 "strip_size_kb": 0, 00:13:50.760 "state": "online", 00:13:50.760 "raid_level": "raid1", 00:13:50.760 "superblock": true, 00:13:50.760 "num_base_bdevs": 4, 00:13:50.760 "num_base_bdevs_discovered": 3, 00:13:50.760 "num_base_bdevs_operational": 3, 00:13:50.760 "base_bdevs_list": [ 00:13:50.760 { 00:13:50.760 "name": null, 00:13:50.760 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:50.760 "is_configured": false, 00:13:50.760 "data_offset": 0, 00:13:50.760 "data_size": 63488 00:13:50.760 }, 00:13:50.760 { 00:13:50.760 "name": "BaseBdev2", 00:13:50.760 "uuid": "a468af4b-33fd-5634-99dc-44fc225c050f", 00:13:50.760 "is_configured": true, 00:13:50.760 "data_offset": 2048, 00:13:50.760 "data_size": 63488 00:13:50.760 }, 00:13:50.760 { 00:13:50.760 "name": "BaseBdev3", 00:13:50.760 "uuid": "bba59b1b-a5ac-5586-a17a-fa7c79f7e077", 00:13:50.760 "is_configured": true, 00:13:50.760 "data_offset": 2048, 00:13:50.760 "data_size": 63488 00:13:50.760 }, 00:13:50.760 { 00:13:50.760 "name": "BaseBdev4", 00:13:50.760 "uuid": "bcf97a61-829b-56dd-ae06-5f30dda1bd20", 00:13:50.760 "is_configured": true, 00:13:50.760 "data_offset": 2048, 00:13:50.760 "data_size": 63488 00:13:50.760 } 00:13:50.760 ] 00:13:50.760 }' 00:13:50.760 15:19:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:50.760 15:19:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:50.760 15:19:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:50.760 15:19:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.760 15:19:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:50.760 [2024-11-19 15:19:40.767768] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:50.760 [2024-11-19 15:19:40.772194] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000c3e420 00:13:50.760 15:19:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.760 15:19:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:13:50.760 [2024-11-19 15:19:40.774159] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:51.701 15:19:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:51.701 15:19:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:51.701 15:19:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:51.701 15:19:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:51.701 15:19:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:51.701 15:19:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:51.701 15:19:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:51.701 15:19:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:51.701 15:19:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:51.701 15:19:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:51.701 15:19:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:51.701 "name": "raid_bdev1", 00:13:51.701 "uuid": "499db5d6-2ee5-45f5-b8aa-305d1e2ff852", 00:13:51.701 "strip_size_kb": 0, 00:13:51.701 "state": "online", 00:13:51.701 "raid_level": "raid1", 00:13:51.701 "superblock": true, 00:13:51.701 "num_base_bdevs": 4, 00:13:51.701 "num_base_bdevs_discovered": 4, 00:13:51.701 "num_base_bdevs_operational": 4, 00:13:51.701 "process": { 00:13:51.701 "type": "rebuild", 00:13:51.701 "target": "spare", 00:13:51.701 "progress": { 00:13:51.701 "blocks": 20480, 00:13:51.701 "percent": 32 00:13:51.701 } 00:13:51.701 }, 00:13:51.701 "base_bdevs_list": [ 00:13:51.701 { 00:13:51.701 "name": "spare", 00:13:51.701 "uuid": "da1587f0-81e2-5332-a1fe-9b88f4b96125", 00:13:51.701 "is_configured": true, 00:13:51.701 "data_offset": 2048, 00:13:51.701 "data_size": 63488 00:13:51.701 }, 00:13:51.701 { 00:13:51.701 "name": "BaseBdev2", 00:13:51.701 "uuid": "a468af4b-33fd-5634-99dc-44fc225c050f", 00:13:51.701 "is_configured": true, 00:13:51.701 "data_offset": 2048, 00:13:51.701 "data_size": 63488 00:13:51.701 }, 00:13:51.701 { 00:13:51.701 "name": "BaseBdev3", 00:13:51.701 "uuid": "bba59b1b-a5ac-5586-a17a-fa7c79f7e077", 00:13:51.701 "is_configured": true, 00:13:51.701 "data_offset": 2048, 00:13:51.701 "data_size": 63488 00:13:51.701 }, 00:13:51.701 { 00:13:51.701 "name": "BaseBdev4", 00:13:51.701 "uuid": "bcf97a61-829b-56dd-ae06-5f30dda1bd20", 00:13:51.701 "is_configured": true, 00:13:51.701 "data_offset": 2048, 00:13:51.701 "data_size": 63488 00:13:51.701 } 00:13:51.701 ] 00:13:51.701 }' 00:13:51.701 15:19:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:51.701 15:19:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:51.701 15:19:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:51.701 15:19:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:51.701 15:19:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:13:51.701 15:19:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:51.701 15:19:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:51.701 [2024-11-19 15:19:41.927080] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:51.701 [2024-11-19 15:19:41.978816] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:51.701 [2024-11-19 15:19:41.978909] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:51.701 [2024-11-19 15:19:41.978931] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:51.701 [2024-11-19 15:19:41.978941] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:51.701 15:19:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:51.701 15:19:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:13:51.701 15:19:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:51.701 15:19:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:51.701 15:19:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:51.701 15:19:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:51.701 15:19:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:51.701 15:19:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:51.701 15:19:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:51.701 15:19:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:51.701 15:19:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:51.701 15:19:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:51.701 15:19:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:51.701 15:19:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:51.701 15:19:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:51.701 15:19:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:51.701 15:19:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:51.701 "name": "raid_bdev1", 00:13:51.701 "uuid": "499db5d6-2ee5-45f5-b8aa-305d1e2ff852", 00:13:51.701 "strip_size_kb": 0, 00:13:51.701 "state": "online", 00:13:51.701 "raid_level": "raid1", 00:13:51.701 "superblock": true, 00:13:51.701 "num_base_bdevs": 4, 00:13:51.701 "num_base_bdevs_discovered": 3, 00:13:51.701 "num_base_bdevs_operational": 3, 00:13:51.701 "base_bdevs_list": [ 00:13:51.701 { 00:13:51.701 "name": null, 00:13:51.701 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:51.701 "is_configured": false, 00:13:51.701 "data_offset": 0, 00:13:51.701 "data_size": 63488 00:13:51.701 }, 00:13:51.701 { 00:13:51.701 "name": "BaseBdev2", 00:13:51.701 "uuid": "a468af4b-33fd-5634-99dc-44fc225c050f", 00:13:51.701 "is_configured": true, 00:13:51.701 "data_offset": 2048, 00:13:51.701 "data_size": 63488 00:13:51.701 }, 00:13:51.701 { 00:13:51.701 "name": "BaseBdev3", 00:13:51.701 "uuid": "bba59b1b-a5ac-5586-a17a-fa7c79f7e077", 00:13:51.701 "is_configured": true, 00:13:51.701 "data_offset": 2048, 00:13:51.701 "data_size": 63488 00:13:51.701 }, 00:13:51.701 { 00:13:51.701 "name": "BaseBdev4", 00:13:51.701 "uuid": "bcf97a61-829b-56dd-ae06-5f30dda1bd20", 00:13:51.701 "is_configured": true, 00:13:51.701 "data_offset": 2048, 00:13:51.701 "data_size": 63488 00:13:51.701 } 00:13:51.701 ] 00:13:51.701 }' 00:13:51.701 15:19:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:51.701 15:19:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:52.273 15:19:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:52.273 15:19:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:52.273 15:19:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:52.273 15:19:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:52.273 15:19:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:52.273 15:19:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:52.273 15:19:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:52.273 15:19:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:52.273 15:19:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:52.273 15:19:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:52.273 15:19:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:52.273 "name": "raid_bdev1", 00:13:52.273 "uuid": "499db5d6-2ee5-45f5-b8aa-305d1e2ff852", 00:13:52.273 "strip_size_kb": 0, 00:13:52.273 "state": "online", 00:13:52.273 "raid_level": "raid1", 00:13:52.273 "superblock": true, 00:13:52.273 "num_base_bdevs": 4, 00:13:52.273 "num_base_bdevs_discovered": 3, 00:13:52.273 "num_base_bdevs_operational": 3, 00:13:52.273 "base_bdevs_list": [ 00:13:52.273 { 00:13:52.273 "name": null, 00:13:52.273 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:52.273 "is_configured": false, 00:13:52.273 "data_offset": 0, 00:13:52.273 "data_size": 63488 00:13:52.273 }, 00:13:52.273 { 00:13:52.273 "name": "BaseBdev2", 00:13:52.273 "uuid": "a468af4b-33fd-5634-99dc-44fc225c050f", 00:13:52.273 "is_configured": true, 00:13:52.273 "data_offset": 2048, 00:13:52.273 "data_size": 63488 00:13:52.273 }, 00:13:52.273 { 00:13:52.273 "name": "BaseBdev3", 00:13:52.273 "uuid": "bba59b1b-a5ac-5586-a17a-fa7c79f7e077", 00:13:52.273 "is_configured": true, 00:13:52.273 "data_offset": 2048, 00:13:52.273 "data_size": 63488 00:13:52.273 }, 00:13:52.273 { 00:13:52.273 "name": "BaseBdev4", 00:13:52.273 "uuid": "bcf97a61-829b-56dd-ae06-5f30dda1bd20", 00:13:52.273 "is_configured": true, 00:13:52.273 "data_offset": 2048, 00:13:52.273 "data_size": 63488 00:13:52.273 } 00:13:52.273 ] 00:13:52.273 }' 00:13:52.273 15:19:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:52.273 15:19:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:52.273 15:19:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:52.273 15:19:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:52.273 15:19:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:52.273 15:19:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:52.273 15:19:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:52.273 [2024-11-19 15:19:42.506643] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:52.273 [2024-11-19 15:19:42.510672] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000c3e4f0 00:13:52.273 15:19:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:52.273 15:19:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:13:52.273 [2024-11-19 15:19:42.512584] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:53.214 15:19:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:53.214 15:19:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:53.214 15:19:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:53.214 15:19:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:53.214 15:19:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:53.214 15:19:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:53.214 15:19:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:53.214 15:19:43 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:53.214 15:19:43 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:53.214 15:19:43 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:53.474 15:19:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:53.474 "name": "raid_bdev1", 00:13:53.474 "uuid": "499db5d6-2ee5-45f5-b8aa-305d1e2ff852", 00:13:53.474 "strip_size_kb": 0, 00:13:53.474 "state": "online", 00:13:53.474 "raid_level": "raid1", 00:13:53.474 "superblock": true, 00:13:53.474 "num_base_bdevs": 4, 00:13:53.474 "num_base_bdevs_discovered": 4, 00:13:53.474 "num_base_bdevs_operational": 4, 00:13:53.475 "process": { 00:13:53.475 "type": "rebuild", 00:13:53.475 "target": "spare", 00:13:53.475 "progress": { 00:13:53.475 "blocks": 20480, 00:13:53.475 "percent": 32 00:13:53.475 } 00:13:53.475 }, 00:13:53.475 "base_bdevs_list": [ 00:13:53.475 { 00:13:53.475 "name": "spare", 00:13:53.475 "uuid": "da1587f0-81e2-5332-a1fe-9b88f4b96125", 00:13:53.475 "is_configured": true, 00:13:53.475 "data_offset": 2048, 00:13:53.475 "data_size": 63488 00:13:53.475 }, 00:13:53.475 { 00:13:53.475 "name": "BaseBdev2", 00:13:53.475 "uuid": "a468af4b-33fd-5634-99dc-44fc225c050f", 00:13:53.475 "is_configured": true, 00:13:53.475 "data_offset": 2048, 00:13:53.475 "data_size": 63488 00:13:53.475 }, 00:13:53.475 { 00:13:53.475 "name": "BaseBdev3", 00:13:53.475 "uuid": "bba59b1b-a5ac-5586-a17a-fa7c79f7e077", 00:13:53.475 "is_configured": true, 00:13:53.475 "data_offset": 2048, 00:13:53.475 "data_size": 63488 00:13:53.475 }, 00:13:53.475 { 00:13:53.475 "name": "BaseBdev4", 00:13:53.475 "uuid": "bcf97a61-829b-56dd-ae06-5f30dda1bd20", 00:13:53.475 "is_configured": true, 00:13:53.475 "data_offset": 2048, 00:13:53.475 "data_size": 63488 00:13:53.475 } 00:13:53.475 ] 00:13:53.475 }' 00:13:53.475 15:19:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:53.475 15:19:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:53.475 15:19:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:53.475 15:19:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:53.475 15:19:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:13:53.475 15:19:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:13:53.475 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:13:53.475 15:19:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:13:53.475 15:19:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:13:53.475 15:19:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:13:53.475 15:19:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:13:53.475 15:19:43 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:53.475 15:19:43 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:53.475 [2024-11-19 15:19:43.669151] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:13:53.735 [2024-11-19 15:19:43.816783] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000c3e4f0 00:13:53.736 15:19:43 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:53.736 15:19:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:13:53.736 15:19:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:13:53.736 15:19:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:53.736 15:19:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:53.736 15:19:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:53.736 15:19:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:53.736 15:19:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:53.736 15:19:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:53.736 15:19:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:53.736 15:19:43 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:53.736 15:19:43 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:53.736 15:19:43 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:53.736 15:19:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:53.736 "name": "raid_bdev1", 00:13:53.736 "uuid": "499db5d6-2ee5-45f5-b8aa-305d1e2ff852", 00:13:53.736 "strip_size_kb": 0, 00:13:53.736 "state": "online", 00:13:53.736 "raid_level": "raid1", 00:13:53.736 "superblock": true, 00:13:53.736 "num_base_bdevs": 4, 00:13:53.736 "num_base_bdevs_discovered": 3, 00:13:53.736 "num_base_bdevs_operational": 3, 00:13:53.736 "process": { 00:13:53.736 "type": "rebuild", 00:13:53.736 "target": "spare", 00:13:53.736 "progress": { 00:13:53.736 "blocks": 24576, 00:13:53.736 "percent": 38 00:13:53.736 } 00:13:53.736 }, 00:13:53.736 "base_bdevs_list": [ 00:13:53.736 { 00:13:53.736 "name": "spare", 00:13:53.736 "uuid": "da1587f0-81e2-5332-a1fe-9b88f4b96125", 00:13:53.736 "is_configured": true, 00:13:53.736 "data_offset": 2048, 00:13:53.736 "data_size": 63488 00:13:53.736 }, 00:13:53.736 { 00:13:53.736 "name": null, 00:13:53.736 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:53.736 "is_configured": false, 00:13:53.736 "data_offset": 0, 00:13:53.736 "data_size": 63488 00:13:53.736 }, 00:13:53.736 { 00:13:53.736 "name": "BaseBdev3", 00:13:53.736 "uuid": "bba59b1b-a5ac-5586-a17a-fa7c79f7e077", 00:13:53.736 "is_configured": true, 00:13:53.736 "data_offset": 2048, 00:13:53.736 "data_size": 63488 00:13:53.736 }, 00:13:53.736 { 00:13:53.736 "name": "BaseBdev4", 00:13:53.736 "uuid": "bcf97a61-829b-56dd-ae06-5f30dda1bd20", 00:13:53.736 "is_configured": true, 00:13:53.736 "data_offset": 2048, 00:13:53.736 "data_size": 63488 00:13:53.736 } 00:13:53.736 ] 00:13:53.736 }' 00:13:53.736 15:19:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:53.736 15:19:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:53.736 15:19:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:53.736 15:19:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:53.736 15:19:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=381 00:13:53.736 15:19:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:53.736 15:19:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:53.736 15:19:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:53.736 15:19:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:53.736 15:19:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:53.736 15:19:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:53.736 15:19:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:53.736 15:19:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:53.736 15:19:43 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:53.736 15:19:43 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:53.736 15:19:43 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:53.736 15:19:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:53.736 "name": "raid_bdev1", 00:13:53.736 "uuid": "499db5d6-2ee5-45f5-b8aa-305d1e2ff852", 00:13:53.736 "strip_size_kb": 0, 00:13:53.736 "state": "online", 00:13:53.736 "raid_level": "raid1", 00:13:53.736 "superblock": true, 00:13:53.736 "num_base_bdevs": 4, 00:13:53.736 "num_base_bdevs_discovered": 3, 00:13:53.736 "num_base_bdevs_operational": 3, 00:13:53.736 "process": { 00:13:53.736 "type": "rebuild", 00:13:53.736 "target": "spare", 00:13:53.736 "progress": { 00:13:53.736 "blocks": 26624, 00:13:53.736 "percent": 41 00:13:53.736 } 00:13:53.736 }, 00:13:53.736 "base_bdevs_list": [ 00:13:53.736 { 00:13:53.736 "name": "spare", 00:13:53.736 "uuid": "da1587f0-81e2-5332-a1fe-9b88f4b96125", 00:13:53.736 "is_configured": true, 00:13:53.736 "data_offset": 2048, 00:13:53.736 "data_size": 63488 00:13:53.736 }, 00:13:53.736 { 00:13:53.736 "name": null, 00:13:53.736 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:53.736 "is_configured": false, 00:13:53.736 "data_offset": 0, 00:13:53.736 "data_size": 63488 00:13:53.736 }, 00:13:53.736 { 00:13:53.736 "name": "BaseBdev3", 00:13:53.736 "uuid": "bba59b1b-a5ac-5586-a17a-fa7c79f7e077", 00:13:53.736 "is_configured": true, 00:13:53.736 "data_offset": 2048, 00:13:53.736 "data_size": 63488 00:13:53.736 }, 00:13:53.736 { 00:13:53.736 "name": "BaseBdev4", 00:13:53.736 "uuid": "bcf97a61-829b-56dd-ae06-5f30dda1bd20", 00:13:53.736 "is_configured": true, 00:13:53.736 "data_offset": 2048, 00:13:53.736 "data_size": 63488 00:13:53.736 } 00:13:53.736 ] 00:13:53.736 }' 00:13:53.736 15:19:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:53.736 15:19:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:53.736 15:19:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:53.996 15:19:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:53.996 15:19:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:54.935 15:19:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:54.935 15:19:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:54.935 15:19:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:54.935 15:19:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:54.935 15:19:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:54.935 15:19:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:54.935 15:19:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:54.935 15:19:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:54.935 15:19:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:54.935 15:19:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:54.935 15:19:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:54.935 15:19:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:54.935 "name": "raid_bdev1", 00:13:54.935 "uuid": "499db5d6-2ee5-45f5-b8aa-305d1e2ff852", 00:13:54.935 "strip_size_kb": 0, 00:13:54.935 "state": "online", 00:13:54.935 "raid_level": "raid1", 00:13:54.935 "superblock": true, 00:13:54.935 "num_base_bdevs": 4, 00:13:54.935 "num_base_bdevs_discovered": 3, 00:13:54.935 "num_base_bdevs_operational": 3, 00:13:54.935 "process": { 00:13:54.935 "type": "rebuild", 00:13:54.935 "target": "spare", 00:13:54.935 "progress": { 00:13:54.935 "blocks": 51200, 00:13:54.935 "percent": 80 00:13:54.935 } 00:13:54.935 }, 00:13:54.935 "base_bdevs_list": [ 00:13:54.935 { 00:13:54.935 "name": "spare", 00:13:54.935 "uuid": "da1587f0-81e2-5332-a1fe-9b88f4b96125", 00:13:54.935 "is_configured": true, 00:13:54.935 "data_offset": 2048, 00:13:54.935 "data_size": 63488 00:13:54.935 }, 00:13:54.935 { 00:13:54.935 "name": null, 00:13:54.935 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:54.935 "is_configured": false, 00:13:54.935 "data_offset": 0, 00:13:54.935 "data_size": 63488 00:13:54.935 }, 00:13:54.935 { 00:13:54.935 "name": "BaseBdev3", 00:13:54.935 "uuid": "bba59b1b-a5ac-5586-a17a-fa7c79f7e077", 00:13:54.935 "is_configured": true, 00:13:54.935 "data_offset": 2048, 00:13:54.935 "data_size": 63488 00:13:54.935 }, 00:13:54.935 { 00:13:54.935 "name": "BaseBdev4", 00:13:54.935 "uuid": "bcf97a61-829b-56dd-ae06-5f30dda1bd20", 00:13:54.935 "is_configured": true, 00:13:54.936 "data_offset": 2048, 00:13:54.936 "data_size": 63488 00:13:54.936 } 00:13:54.936 ] 00:13:54.936 }' 00:13:54.936 15:19:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:54.936 15:19:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:54.936 15:19:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:54.936 15:19:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:54.936 15:19:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:55.506 [2024-11-19 15:19:45.723710] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:13:55.506 [2024-11-19 15:19:45.723821] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:13:55.506 [2024-11-19 15:19:45.723927] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:56.076 15:19:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:56.076 15:19:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:56.076 15:19:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:56.076 15:19:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:56.076 15:19:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:56.076 15:19:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:56.076 15:19:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:56.076 15:19:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:56.076 15:19:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:56.076 15:19:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:56.076 15:19:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:56.076 15:19:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:56.076 "name": "raid_bdev1", 00:13:56.076 "uuid": "499db5d6-2ee5-45f5-b8aa-305d1e2ff852", 00:13:56.076 "strip_size_kb": 0, 00:13:56.076 "state": "online", 00:13:56.076 "raid_level": "raid1", 00:13:56.076 "superblock": true, 00:13:56.076 "num_base_bdevs": 4, 00:13:56.076 "num_base_bdevs_discovered": 3, 00:13:56.076 "num_base_bdevs_operational": 3, 00:13:56.076 "base_bdevs_list": [ 00:13:56.076 { 00:13:56.076 "name": "spare", 00:13:56.076 "uuid": "da1587f0-81e2-5332-a1fe-9b88f4b96125", 00:13:56.076 "is_configured": true, 00:13:56.076 "data_offset": 2048, 00:13:56.076 "data_size": 63488 00:13:56.076 }, 00:13:56.076 { 00:13:56.076 "name": null, 00:13:56.076 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:56.076 "is_configured": false, 00:13:56.076 "data_offset": 0, 00:13:56.076 "data_size": 63488 00:13:56.076 }, 00:13:56.076 { 00:13:56.076 "name": "BaseBdev3", 00:13:56.076 "uuid": "bba59b1b-a5ac-5586-a17a-fa7c79f7e077", 00:13:56.076 "is_configured": true, 00:13:56.076 "data_offset": 2048, 00:13:56.076 "data_size": 63488 00:13:56.076 }, 00:13:56.076 { 00:13:56.076 "name": "BaseBdev4", 00:13:56.076 "uuid": "bcf97a61-829b-56dd-ae06-5f30dda1bd20", 00:13:56.076 "is_configured": true, 00:13:56.076 "data_offset": 2048, 00:13:56.076 "data_size": 63488 00:13:56.076 } 00:13:56.076 ] 00:13:56.076 }' 00:13:56.076 15:19:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:56.076 15:19:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:13:56.076 15:19:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:56.337 15:19:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:13:56.337 15:19:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:13:56.337 15:19:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:56.337 15:19:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:56.337 15:19:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:56.337 15:19:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:56.337 15:19:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:56.337 15:19:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:56.337 15:19:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:56.337 15:19:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:56.337 15:19:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:56.337 15:19:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:56.337 15:19:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:56.337 "name": "raid_bdev1", 00:13:56.337 "uuid": "499db5d6-2ee5-45f5-b8aa-305d1e2ff852", 00:13:56.337 "strip_size_kb": 0, 00:13:56.337 "state": "online", 00:13:56.337 "raid_level": "raid1", 00:13:56.337 "superblock": true, 00:13:56.337 "num_base_bdevs": 4, 00:13:56.337 "num_base_bdevs_discovered": 3, 00:13:56.337 "num_base_bdevs_operational": 3, 00:13:56.337 "base_bdevs_list": [ 00:13:56.337 { 00:13:56.337 "name": "spare", 00:13:56.337 "uuid": "da1587f0-81e2-5332-a1fe-9b88f4b96125", 00:13:56.337 "is_configured": true, 00:13:56.337 "data_offset": 2048, 00:13:56.337 "data_size": 63488 00:13:56.337 }, 00:13:56.337 { 00:13:56.337 "name": null, 00:13:56.337 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:56.337 "is_configured": false, 00:13:56.337 "data_offset": 0, 00:13:56.337 "data_size": 63488 00:13:56.337 }, 00:13:56.337 { 00:13:56.337 "name": "BaseBdev3", 00:13:56.337 "uuid": "bba59b1b-a5ac-5586-a17a-fa7c79f7e077", 00:13:56.337 "is_configured": true, 00:13:56.337 "data_offset": 2048, 00:13:56.337 "data_size": 63488 00:13:56.337 }, 00:13:56.337 { 00:13:56.337 "name": "BaseBdev4", 00:13:56.337 "uuid": "bcf97a61-829b-56dd-ae06-5f30dda1bd20", 00:13:56.337 "is_configured": true, 00:13:56.337 "data_offset": 2048, 00:13:56.337 "data_size": 63488 00:13:56.337 } 00:13:56.337 ] 00:13:56.337 }' 00:13:56.337 15:19:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:56.337 15:19:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:56.337 15:19:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:56.337 15:19:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:56.337 15:19:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:13:56.337 15:19:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:56.337 15:19:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:56.337 15:19:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:56.337 15:19:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:56.337 15:19:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:56.337 15:19:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:56.337 15:19:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:56.337 15:19:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:56.337 15:19:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:56.337 15:19:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:56.337 15:19:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:56.337 15:19:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:56.337 15:19:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:56.337 15:19:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:56.337 15:19:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:56.337 "name": "raid_bdev1", 00:13:56.337 "uuid": "499db5d6-2ee5-45f5-b8aa-305d1e2ff852", 00:13:56.337 "strip_size_kb": 0, 00:13:56.337 "state": "online", 00:13:56.337 "raid_level": "raid1", 00:13:56.337 "superblock": true, 00:13:56.337 "num_base_bdevs": 4, 00:13:56.337 "num_base_bdevs_discovered": 3, 00:13:56.337 "num_base_bdevs_operational": 3, 00:13:56.337 "base_bdevs_list": [ 00:13:56.337 { 00:13:56.337 "name": "spare", 00:13:56.337 "uuid": "da1587f0-81e2-5332-a1fe-9b88f4b96125", 00:13:56.337 "is_configured": true, 00:13:56.337 "data_offset": 2048, 00:13:56.337 "data_size": 63488 00:13:56.337 }, 00:13:56.337 { 00:13:56.337 "name": null, 00:13:56.337 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:56.337 "is_configured": false, 00:13:56.337 "data_offset": 0, 00:13:56.337 "data_size": 63488 00:13:56.337 }, 00:13:56.337 { 00:13:56.337 "name": "BaseBdev3", 00:13:56.337 "uuid": "bba59b1b-a5ac-5586-a17a-fa7c79f7e077", 00:13:56.337 "is_configured": true, 00:13:56.337 "data_offset": 2048, 00:13:56.337 "data_size": 63488 00:13:56.337 }, 00:13:56.337 { 00:13:56.337 "name": "BaseBdev4", 00:13:56.338 "uuid": "bcf97a61-829b-56dd-ae06-5f30dda1bd20", 00:13:56.338 "is_configured": true, 00:13:56.338 "data_offset": 2048, 00:13:56.338 "data_size": 63488 00:13:56.338 } 00:13:56.338 ] 00:13:56.338 }' 00:13:56.338 15:19:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:56.338 15:19:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:56.912 15:19:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:13:56.912 15:19:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:56.912 15:19:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:56.912 [2024-11-19 15:19:46.953975] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:56.912 [2024-11-19 15:19:46.954025] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:56.912 [2024-11-19 15:19:46.954147] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:56.912 [2024-11-19 15:19:46.954233] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:56.912 [2024-11-19 15:19:46.954255] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:13:56.912 15:19:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:56.912 15:19:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:13:56.912 15:19:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:56.912 15:19:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:56.912 15:19:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:56.912 15:19:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:56.912 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:13:56.912 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:13:56.912 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:13:56.912 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:13:56.912 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:56.912 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:13:56.912 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:56.912 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:13:56.912 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:56.912 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:13:56.912 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:56.912 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:13:56.912 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:13:56.912 /dev/nbd0 00:13:56.912 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:13:56.912 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:13:56.912 15:19:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:13:56.912 15:19:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:13:56.912 15:19:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:56.912 15:19:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:56.912 15:19:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:13:56.912 15:19:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:13:56.912 15:19:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:56.912 15:19:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:56.912 15:19:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:56.912 1+0 records in 00:13:56.912 1+0 records out 00:13:56.912 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000356063 s, 11.5 MB/s 00:13:56.912 15:19:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:56.912 15:19:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:13:56.912 15:19:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:56.912 15:19:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:56.912 15:19:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:13:56.912 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:56.912 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:13:56.912 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:13:57.186 /dev/nbd1 00:13:57.186 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:13:57.186 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:13:57.186 15:19:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:13:57.187 15:19:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:13:57.187 15:19:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:57.187 15:19:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:57.187 15:19:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:13:57.187 15:19:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:13:57.187 15:19:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:57.187 15:19:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:57.187 15:19:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:57.187 1+0 records in 00:13:57.187 1+0 records out 00:13:57.187 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000394217 s, 10.4 MB/s 00:13:57.187 15:19:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:57.187 15:19:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:13:57.187 15:19:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:57.187 15:19:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:57.187 15:19:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:13:57.187 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:57.187 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:13:57.187 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:13:57.462 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:13:57.462 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:57.462 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:13:57.462 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:57.462 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:13:57.462 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:57.462 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:13:57.462 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:13:57.462 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:13:57.462 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:13:57.462 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:57.462 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:57.462 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:13:57.462 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:13:57.462 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:13:57.462 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:57.462 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:13:57.721 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:13:57.721 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:13:57.721 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:13:57.721 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:57.721 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:57.721 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:13:57.721 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:13:57.721 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:13:57.721 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:13:57.721 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:13:57.721 15:19:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:57.721 15:19:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:57.721 15:19:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:57.721 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:57.721 15:19:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:57.721 15:19:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:57.721 [2024-11-19 15:19:47.995045] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:57.721 [2024-11-19 15:19:47.995123] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:57.721 [2024-11-19 15:19:47.995149] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:13:57.721 [2024-11-19 15:19:47.995166] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:57.721 [2024-11-19 15:19:47.997430] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:57.721 [2024-11-19 15:19:47.997479] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:57.721 [2024-11-19 15:19:47.997570] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:13:57.721 [2024-11-19 15:19:47.997620] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:57.721 [2024-11-19 15:19:47.997736] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:57.721 [2024-11-19 15:19:47.997868] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:13:57.721 spare 00:13:57.721 15:19:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:57.721 15:19:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:13:57.721 15:19:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:57.721 15:19:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:57.981 [2024-11-19 15:19:48.097775] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:13:57.981 [2024-11-19 15:19:48.097811] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:13:57.981 [2024-11-19 15:19:48.098154] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000caeb00 00:13:57.981 [2024-11-19 15:19:48.098341] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:13:57.981 [2024-11-19 15:19:48.098352] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:13:57.981 [2024-11-19 15:19:48.098510] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:57.981 15:19:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:57.981 15:19:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:13:57.981 15:19:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:57.981 15:19:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:57.981 15:19:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:57.981 15:19:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:57.981 15:19:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:57.981 15:19:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:57.981 15:19:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:57.981 15:19:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:57.981 15:19:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:57.981 15:19:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:57.981 15:19:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:57.981 15:19:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:57.981 15:19:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:57.981 15:19:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:57.981 15:19:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:57.981 "name": "raid_bdev1", 00:13:57.981 "uuid": "499db5d6-2ee5-45f5-b8aa-305d1e2ff852", 00:13:57.981 "strip_size_kb": 0, 00:13:57.981 "state": "online", 00:13:57.981 "raid_level": "raid1", 00:13:57.981 "superblock": true, 00:13:57.981 "num_base_bdevs": 4, 00:13:57.981 "num_base_bdevs_discovered": 3, 00:13:57.981 "num_base_bdevs_operational": 3, 00:13:57.981 "base_bdevs_list": [ 00:13:57.981 { 00:13:57.981 "name": "spare", 00:13:57.981 "uuid": "da1587f0-81e2-5332-a1fe-9b88f4b96125", 00:13:57.981 "is_configured": true, 00:13:57.981 "data_offset": 2048, 00:13:57.981 "data_size": 63488 00:13:57.981 }, 00:13:57.981 { 00:13:57.981 "name": null, 00:13:57.981 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:57.981 "is_configured": false, 00:13:57.981 "data_offset": 2048, 00:13:57.981 "data_size": 63488 00:13:57.981 }, 00:13:57.981 { 00:13:57.981 "name": "BaseBdev3", 00:13:57.981 "uuid": "bba59b1b-a5ac-5586-a17a-fa7c79f7e077", 00:13:57.981 "is_configured": true, 00:13:57.981 "data_offset": 2048, 00:13:57.981 "data_size": 63488 00:13:57.981 }, 00:13:57.981 { 00:13:57.981 "name": "BaseBdev4", 00:13:57.981 "uuid": "bcf97a61-829b-56dd-ae06-5f30dda1bd20", 00:13:57.981 "is_configured": true, 00:13:57.981 "data_offset": 2048, 00:13:57.981 "data_size": 63488 00:13:57.981 } 00:13:57.981 ] 00:13:57.981 }' 00:13:57.981 15:19:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:57.981 15:19:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:58.241 15:19:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:58.241 15:19:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:58.241 15:19:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:58.241 15:19:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:58.241 15:19:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:58.241 15:19:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:58.241 15:19:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:58.241 15:19:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:58.241 15:19:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:58.241 15:19:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:58.500 15:19:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:58.500 "name": "raid_bdev1", 00:13:58.500 "uuid": "499db5d6-2ee5-45f5-b8aa-305d1e2ff852", 00:13:58.500 "strip_size_kb": 0, 00:13:58.500 "state": "online", 00:13:58.500 "raid_level": "raid1", 00:13:58.500 "superblock": true, 00:13:58.500 "num_base_bdevs": 4, 00:13:58.500 "num_base_bdevs_discovered": 3, 00:13:58.500 "num_base_bdevs_operational": 3, 00:13:58.500 "base_bdevs_list": [ 00:13:58.500 { 00:13:58.500 "name": "spare", 00:13:58.500 "uuid": "da1587f0-81e2-5332-a1fe-9b88f4b96125", 00:13:58.500 "is_configured": true, 00:13:58.500 "data_offset": 2048, 00:13:58.500 "data_size": 63488 00:13:58.500 }, 00:13:58.500 { 00:13:58.500 "name": null, 00:13:58.500 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:58.500 "is_configured": false, 00:13:58.500 "data_offset": 2048, 00:13:58.500 "data_size": 63488 00:13:58.500 }, 00:13:58.500 { 00:13:58.500 "name": "BaseBdev3", 00:13:58.500 "uuid": "bba59b1b-a5ac-5586-a17a-fa7c79f7e077", 00:13:58.500 "is_configured": true, 00:13:58.500 "data_offset": 2048, 00:13:58.500 "data_size": 63488 00:13:58.500 }, 00:13:58.500 { 00:13:58.500 "name": "BaseBdev4", 00:13:58.500 "uuid": "bcf97a61-829b-56dd-ae06-5f30dda1bd20", 00:13:58.500 "is_configured": true, 00:13:58.500 "data_offset": 2048, 00:13:58.500 "data_size": 63488 00:13:58.500 } 00:13:58.500 ] 00:13:58.500 }' 00:13:58.500 15:19:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:58.500 15:19:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:58.500 15:19:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:58.500 15:19:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:58.501 15:19:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:58.501 15:19:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:13:58.501 15:19:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:58.501 15:19:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:58.501 15:19:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:58.501 15:19:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:13:58.501 15:19:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:13:58.501 15:19:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:58.501 15:19:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:58.501 [2024-11-19 15:19:48.709824] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:58.501 15:19:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:58.501 15:19:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:58.501 15:19:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:58.501 15:19:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:58.501 15:19:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:58.501 15:19:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:58.501 15:19:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:58.501 15:19:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:58.501 15:19:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:58.501 15:19:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:58.501 15:19:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:58.501 15:19:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:58.501 15:19:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:58.501 15:19:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:58.501 15:19:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:58.501 15:19:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:58.501 15:19:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:58.501 "name": "raid_bdev1", 00:13:58.501 "uuid": "499db5d6-2ee5-45f5-b8aa-305d1e2ff852", 00:13:58.501 "strip_size_kb": 0, 00:13:58.501 "state": "online", 00:13:58.501 "raid_level": "raid1", 00:13:58.501 "superblock": true, 00:13:58.501 "num_base_bdevs": 4, 00:13:58.501 "num_base_bdevs_discovered": 2, 00:13:58.501 "num_base_bdevs_operational": 2, 00:13:58.501 "base_bdevs_list": [ 00:13:58.501 { 00:13:58.501 "name": null, 00:13:58.501 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:58.501 "is_configured": false, 00:13:58.501 "data_offset": 0, 00:13:58.501 "data_size": 63488 00:13:58.501 }, 00:13:58.501 { 00:13:58.501 "name": null, 00:13:58.501 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:58.501 "is_configured": false, 00:13:58.501 "data_offset": 2048, 00:13:58.501 "data_size": 63488 00:13:58.501 }, 00:13:58.501 { 00:13:58.501 "name": "BaseBdev3", 00:13:58.501 "uuid": "bba59b1b-a5ac-5586-a17a-fa7c79f7e077", 00:13:58.501 "is_configured": true, 00:13:58.501 "data_offset": 2048, 00:13:58.501 "data_size": 63488 00:13:58.501 }, 00:13:58.501 { 00:13:58.501 "name": "BaseBdev4", 00:13:58.501 "uuid": "bcf97a61-829b-56dd-ae06-5f30dda1bd20", 00:13:58.501 "is_configured": true, 00:13:58.501 "data_offset": 2048, 00:13:58.501 "data_size": 63488 00:13:58.501 } 00:13:58.501 ] 00:13:58.501 }' 00:13:58.501 15:19:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:58.501 15:19:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.070 15:19:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:59.070 15:19:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:59.070 15:19:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.070 [2024-11-19 15:19:49.121108] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:59.070 [2024-11-19 15:19:49.121297] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:13:59.070 [2024-11-19 15:19:49.121325] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:13:59.070 [2024-11-19 15:19:49.121362] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:59.070 [2024-11-19 15:19:49.125468] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000caebd0 00:13:59.070 15:19:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:59.070 15:19:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:13:59.070 [2024-11-19 15:19:49.127350] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:00.009 15:19:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:00.009 15:19:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:00.009 15:19:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:00.009 15:19:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:00.009 15:19:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:00.009 15:19:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:00.009 15:19:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:00.009 15:19:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:00.009 15:19:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:00.009 15:19:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:00.009 15:19:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:00.009 "name": "raid_bdev1", 00:14:00.009 "uuid": "499db5d6-2ee5-45f5-b8aa-305d1e2ff852", 00:14:00.009 "strip_size_kb": 0, 00:14:00.009 "state": "online", 00:14:00.009 "raid_level": "raid1", 00:14:00.009 "superblock": true, 00:14:00.009 "num_base_bdevs": 4, 00:14:00.009 "num_base_bdevs_discovered": 3, 00:14:00.009 "num_base_bdevs_operational": 3, 00:14:00.009 "process": { 00:14:00.009 "type": "rebuild", 00:14:00.009 "target": "spare", 00:14:00.009 "progress": { 00:14:00.009 "blocks": 20480, 00:14:00.009 "percent": 32 00:14:00.009 } 00:14:00.009 }, 00:14:00.009 "base_bdevs_list": [ 00:14:00.009 { 00:14:00.009 "name": "spare", 00:14:00.009 "uuid": "da1587f0-81e2-5332-a1fe-9b88f4b96125", 00:14:00.009 "is_configured": true, 00:14:00.009 "data_offset": 2048, 00:14:00.009 "data_size": 63488 00:14:00.009 }, 00:14:00.009 { 00:14:00.009 "name": null, 00:14:00.009 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:00.009 "is_configured": false, 00:14:00.009 "data_offset": 2048, 00:14:00.009 "data_size": 63488 00:14:00.009 }, 00:14:00.009 { 00:14:00.009 "name": "BaseBdev3", 00:14:00.009 "uuid": "bba59b1b-a5ac-5586-a17a-fa7c79f7e077", 00:14:00.009 "is_configured": true, 00:14:00.009 "data_offset": 2048, 00:14:00.009 "data_size": 63488 00:14:00.009 }, 00:14:00.009 { 00:14:00.009 "name": "BaseBdev4", 00:14:00.009 "uuid": "bcf97a61-829b-56dd-ae06-5f30dda1bd20", 00:14:00.009 "is_configured": true, 00:14:00.009 "data_offset": 2048, 00:14:00.009 "data_size": 63488 00:14:00.009 } 00:14:00.009 ] 00:14:00.009 }' 00:14:00.009 15:19:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:00.009 15:19:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:00.009 15:19:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:00.009 15:19:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:00.009 15:19:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:14:00.009 15:19:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:00.009 15:19:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:00.009 [2024-11-19 15:19:50.288474] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:00.009 [2024-11-19 15:19:50.331383] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:14:00.009 [2024-11-19 15:19:50.331470] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:00.009 [2024-11-19 15:19:50.331488] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:00.009 [2024-11-19 15:19:50.331499] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:14:00.009 15:19:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:00.009 15:19:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:14:00.009 15:19:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:00.009 15:19:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:00.009 15:19:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:00.009 15:19:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:00.009 15:19:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:00.009 15:19:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:00.009 15:19:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:00.009 15:19:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:00.009 15:19:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:00.269 15:19:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:00.269 15:19:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:00.269 15:19:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:00.269 15:19:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:00.269 15:19:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:00.269 15:19:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:00.269 "name": "raid_bdev1", 00:14:00.269 "uuid": "499db5d6-2ee5-45f5-b8aa-305d1e2ff852", 00:14:00.269 "strip_size_kb": 0, 00:14:00.269 "state": "online", 00:14:00.269 "raid_level": "raid1", 00:14:00.269 "superblock": true, 00:14:00.269 "num_base_bdevs": 4, 00:14:00.269 "num_base_bdevs_discovered": 2, 00:14:00.269 "num_base_bdevs_operational": 2, 00:14:00.269 "base_bdevs_list": [ 00:14:00.269 { 00:14:00.269 "name": null, 00:14:00.269 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:00.269 "is_configured": false, 00:14:00.269 "data_offset": 0, 00:14:00.269 "data_size": 63488 00:14:00.269 }, 00:14:00.269 { 00:14:00.269 "name": null, 00:14:00.269 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:00.269 "is_configured": false, 00:14:00.269 "data_offset": 2048, 00:14:00.269 "data_size": 63488 00:14:00.269 }, 00:14:00.269 { 00:14:00.269 "name": "BaseBdev3", 00:14:00.269 "uuid": "bba59b1b-a5ac-5586-a17a-fa7c79f7e077", 00:14:00.269 "is_configured": true, 00:14:00.269 "data_offset": 2048, 00:14:00.269 "data_size": 63488 00:14:00.269 }, 00:14:00.269 { 00:14:00.269 "name": "BaseBdev4", 00:14:00.269 "uuid": "bcf97a61-829b-56dd-ae06-5f30dda1bd20", 00:14:00.269 "is_configured": true, 00:14:00.269 "data_offset": 2048, 00:14:00.269 "data_size": 63488 00:14:00.269 } 00:14:00.269 ] 00:14:00.269 }' 00:14:00.269 15:19:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:00.269 15:19:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:00.528 15:19:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:14:00.528 15:19:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:00.528 15:19:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:00.529 [2024-11-19 15:19:50.771074] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:14:00.529 [2024-11-19 15:19:50.771144] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:00.529 [2024-11-19 15:19:50.771177] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b780 00:14:00.529 [2024-11-19 15:19:50.771194] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:00.529 [2024-11-19 15:19:50.771638] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:00.529 [2024-11-19 15:19:50.771670] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:14:00.529 [2024-11-19 15:19:50.771772] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:14:00.529 [2024-11-19 15:19:50.771800] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:14:00.529 [2024-11-19 15:19:50.771811] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:14:00.529 [2024-11-19 15:19:50.771839] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:00.529 [2024-11-19 15:19:50.775674] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000caeca0 00:14:00.529 spare 00:14:00.529 15:19:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:00.529 15:19:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:14:00.529 [2024-11-19 15:19:50.777550] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:01.469 15:19:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:01.469 15:19:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:01.469 15:19:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:01.469 15:19:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:01.469 15:19:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:01.469 15:19:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:01.469 15:19:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:01.469 15:19:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:01.469 15:19:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:01.729 15:19:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:01.729 15:19:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:01.729 "name": "raid_bdev1", 00:14:01.729 "uuid": "499db5d6-2ee5-45f5-b8aa-305d1e2ff852", 00:14:01.729 "strip_size_kb": 0, 00:14:01.729 "state": "online", 00:14:01.729 "raid_level": "raid1", 00:14:01.729 "superblock": true, 00:14:01.729 "num_base_bdevs": 4, 00:14:01.729 "num_base_bdevs_discovered": 3, 00:14:01.729 "num_base_bdevs_operational": 3, 00:14:01.729 "process": { 00:14:01.729 "type": "rebuild", 00:14:01.729 "target": "spare", 00:14:01.729 "progress": { 00:14:01.729 "blocks": 20480, 00:14:01.729 "percent": 32 00:14:01.729 } 00:14:01.729 }, 00:14:01.729 "base_bdevs_list": [ 00:14:01.729 { 00:14:01.729 "name": "spare", 00:14:01.729 "uuid": "da1587f0-81e2-5332-a1fe-9b88f4b96125", 00:14:01.729 "is_configured": true, 00:14:01.729 "data_offset": 2048, 00:14:01.729 "data_size": 63488 00:14:01.729 }, 00:14:01.729 { 00:14:01.729 "name": null, 00:14:01.729 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:01.729 "is_configured": false, 00:14:01.729 "data_offset": 2048, 00:14:01.729 "data_size": 63488 00:14:01.729 }, 00:14:01.729 { 00:14:01.729 "name": "BaseBdev3", 00:14:01.729 "uuid": "bba59b1b-a5ac-5586-a17a-fa7c79f7e077", 00:14:01.729 "is_configured": true, 00:14:01.729 "data_offset": 2048, 00:14:01.729 "data_size": 63488 00:14:01.729 }, 00:14:01.729 { 00:14:01.729 "name": "BaseBdev4", 00:14:01.729 "uuid": "bcf97a61-829b-56dd-ae06-5f30dda1bd20", 00:14:01.729 "is_configured": true, 00:14:01.729 "data_offset": 2048, 00:14:01.729 "data_size": 63488 00:14:01.729 } 00:14:01.729 ] 00:14:01.729 }' 00:14:01.729 15:19:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:01.729 15:19:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:01.729 15:19:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:01.729 15:19:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:01.729 15:19:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:14:01.729 15:19:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:01.729 15:19:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:01.729 [2024-11-19 15:19:51.914485] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:01.729 [2024-11-19 15:19:51.981618] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:14:01.729 [2024-11-19 15:19:51.981680] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:01.729 [2024-11-19 15:19:51.981717] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:01.729 [2024-11-19 15:19:51.981726] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:14:01.729 15:19:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:01.729 15:19:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:14:01.729 15:19:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:01.729 15:19:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:01.729 15:19:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:01.729 15:19:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:01.729 15:19:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:01.729 15:19:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:01.729 15:19:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:01.729 15:19:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:01.729 15:19:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:01.729 15:19:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:01.729 15:19:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:01.729 15:19:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:01.729 15:19:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:01.730 15:19:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:01.730 15:19:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:01.730 "name": "raid_bdev1", 00:14:01.730 "uuid": "499db5d6-2ee5-45f5-b8aa-305d1e2ff852", 00:14:01.730 "strip_size_kb": 0, 00:14:01.730 "state": "online", 00:14:01.730 "raid_level": "raid1", 00:14:01.730 "superblock": true, 00:14:01.730 "num_base_bdevs": 4, 00:14:01.730 "num_base_bdevs_discovered": 2, 00:14:01.730 "num_base_bdevs_operational": 2, 00:14:01.730 "base_bdevs_list": [ 00:14:01.730 { 00:14:01.730 "name": null, 00:14:01.730 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:01.730 "is_configured": false, 00:14:01.730 "data_offset": 0, 00:14:01.730 "data_size": 63488 00:14:01.730 }, 00:14:01.730 { 00:14:01.730 "name": null, 00:14:01.730 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:01.730 "is_configured": false, 00:14:01.730 "data_offset": 2048, 00:14:01.730 "data_size": 63488 00:14:01.730 }, 00:14:01.730 { 00:14:01.730 "name": "BaseBdev3", 00:14:01.730 "uuid": "bba59b1b-a5ac-5586-a17a-fa7c79f7e077", 00:14:01.730 "is_configured": true, 00:14:01.730 "data_offset": 2048, 00:14:01.730 "data_size": 63488 00:14:01.730 }, 00:14:01.730 { 00:14:01.730 "name": "BaseBdev4", 00:14:01.730 "uuid": "bcf97a61-829b-56dd-ae06-5f30dda1bd20", 00:14:01.730 "is_configured": true, 00:14:01.730 "data_offset": 2048, 00:14:01.730 "data_size": 63488 00:14:01.730 } 00:14:01.730 ] 00:14:01.730 }' 00:14:01.730 15:19:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:01.730 15:19:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:02.300 15:19:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:02.300 15:19:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:02.300 15:19:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:02.300 15:19:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:02.300 15:19:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:02.300 15:19:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:02.300 15:19:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:02.300 15:19:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:02.300 15:19:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:02.300 15:19:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:02.300 15:19:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:02.300 "name": "raid_bdev1", 00:14:02.300 "uuid": "499db5d6-2ee5-45f5-b8aa-305d1e2ff852", 00:14:02.300 "strip_size_kb": 0, 00:14:02.300 "state": "online", 00:14:02.300 "raid_level": "raid1", 00:14:02.300 "superblock": true, 00:14:02.300 "num_base_bdevs": 4, 00:14:02.300 "num_base_bdevs_discovered": 2, 00:14:02.300 "num_base_bdevs_operational": 2, 00:14:02.300 "base_bdevs_list": [ 00:14:02.300 { 00:14:02.300 "name": null, 00:14:02.300 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:02.300 "is_configured": false, 00:14:02.300 "data_offset": 0, 00:14:02.300 "data_size": 63488 00:14:02.300 }, 00:14:02.300 { 00:14:02.300 "name": null, 00:14:02.300 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:02.300 "is_configured": false, 00:14:02.300 "data_offset": 2048, 00:14:02.300 "data_size": 63488 00:14:02.300 }, 00:14:02.300 { 00:14:02.300 "name": "BaseBdev3", 00:14:02.300 "uuid": "bba59b1b-a5ac-5586-a17a-fa7c79f7e077", 00:14:02.300 "is_configured": true, 00:14:02.300 "data_offset": 2048, 00:14:02.300 "data_size": 63488 00:14:02.300 }, 00:14:02.300 { 00:14:02.300 "name": "BaseBdev4", 00:14:02.300 "uuid": "bcf97a61-829b-56dd-ae06-5f30dda1bd20", 00:14:02.300 "is_configured": true, 00:14:02.300 "data_offset": 2048, 00:14:02.300 "data_size": 63488 00:14:02.300 } 00:14:02.300 ] 00:14:02.300 }' 00:14:02.300 15:19:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:02.300 15:19:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:02.300 15:19:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:02.300 15:19:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:02.300 15:19:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:14:02.300 15:19:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:02.300 15:19:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:02.300 15:19:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:02.300 15:19:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:14:02.300 15:19:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:02.300 15:19:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:02.300 [2024-11-19 15:19:52.592833] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:14:02.300 [2024-11-19 15:19:52.592897] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:02.300 [2024-11-19 15:19:52.592923] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000bd80 00:14:02.300 [2024-11-19 15:19:52.592934] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:02.300 [2024-11-19 15:19:52.593363] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:02.300 [2024-11-19 15:19:52.593392] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:14:02.300 [2024-11-19 15:19:52.593473] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:14:02.300 [2024-11-19 15:19:52.593488] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:14:02.300 [2024-11-19 15:19:52.593511] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:14:02.300 [2024-11-19 15:19:52.593522] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:14:02.300 BaseBdev1 00:14:02.300 15:19:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:02.300 15:19:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:14:03.682 15:19:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:14:03.682 15:19:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:03.682 15:19:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:03.682 15:19:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:03.682 15:19:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:03.682 15:19:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:03.682 15:19:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:03.682 15:19:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:03.682 15:19:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:03.682 15:19:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:03.682 15:19:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:03.682 15:19:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:03.682 15:19:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:03.682 15:19:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:03.682 15:19:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:03.682 15:19:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:03.682 "name": "raid_bdev1", 00:14:03.682 "uuid": "499db5d6-2ee5-45f5-b8aa-305d1e2ff852", 00:14:03.682 "strip_size_kb": 0, 00:14:03.682 "state": "online", 00:14:03.682 "raid_level": "raid1", 00:14:03.682 "superblock": true, 00:14:03.682 "num_base_bdevs": 4, 00:14:03.682 "num_base_bdevs_discovered": 2, 00:14:03.682 "num_base_bdevs_operational": 2, 00:14:03.682 "base_bdevs_list": [ 00:14:03.682 { 00:14:03.682 "name": null, 00:14:03.682 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:03.682 "is_configured": false, 00:14:03.682 "data_offset": 0, 00:14:03.682 "data_size": 63488 00:14:03.682 }, 00:14:03.682 { 00:14:03.682 "name": null, 00:14:03.682 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:03.682 "is_configured": false, 00:14:03.682 "data_offset": 2048, 00:14:03.682 "data_size": 63488 00:14:03.682 }, 00:14:03.682 { 00:14:03.682 "name": "BaseBdev3", 00:14:03.682 "uuid": "bba59b1b-a5ac-5586-a17a-fa7c79f7e077", 00:14:03.682 "is_configured": true, 00:14:03.682 "data_offset": 2048, 00:14:03.682 "data_size": 63488 00:14:03.682 }, 00:14:03.682 { 00:14:03.682 "name": "BaseBdev4", 00:14:03.682 "uuid": "bcf97a61-829b-56dd-ae06-5f30dda1bd20", 00:14:03.682 "is_configured": true, 00:14:03.682 "data_offset": 2048, 00:14:03.682 "data_size": 63488 00:14:03.682 } 00:14:03.682 ] 00:14:03.682 }' 00:14:03.682 15:19:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:03.682 15:19:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:03.942 15:19:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:03.942 15:19:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:03.942 15:19:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:03.942 15:19:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:03.942 15:19:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:03.942 15:19:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:03.942 15:19:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:03.942 15:19:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:03.942 15:19:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:03.942 15:19:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:03.942 15:19:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:03.942 "name": "raid_bdev1", 00:14:03.942 "uuid": "499db5d6-2ee5-45f5-b8aa-305d1e2ff852", 00:14:03.942 "strip_size_kb": 0, 00:14:03.942 "state": "online", 00:14:03.942 "raid_level": "raid1", 00:14:03.942 "superblock": true, 00:14:03.942 "num_base_bdevs": 4, 00:14:03.942 "num_base_bdevs_discovered": 2, 00:14:03.942 "num_base_bdevs_operational": 2, 00:14:03.942 "base_bdevs_list": [ 00:14:03.942 { 00:14:03.942 "name": null, 00:14:03.942 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:03.942 "is_configured": false, 00:14:03.942 "data_offset": 0, 00:14:03.942 "data_size": 63488 00:14:03.942 }, 00:14:03.942 { 00:14:03.942 "name": null, 00:14:03.942 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:03.942 "is_configured": false, 00:14:03.942 "data_offset": 2048, 00:14:03.942 "data_size": 63488 00:14:03.942 }, 00:14:03.942 { 00:14:03.942 "name": "BaseBdev3", 00:14:03.942 "uuid": "bba59b1b-a5ac-5586-a17a-fa7c79f7e077", 00:14:03.942 "is_configured": true, 00:14:03.942 "data_offset": 2048, 00:14:03.942 "data_size": 63488 00:14:03.942 }, 00:14:03.942 { 00:14:03.942 "name": "BaseBdev4", 00:14:03.942 "uuid": "bcf97a61-829b-56dd-ae06-5f30dda1bd20", 00:14:03.942 "is_configured": true, 00:14:03.942 "data_offset": 2048, 00:14:03.942 "data_size": 63488 00:14:03.942 } 00:14:03.942 ] 00:14:03.942 }' 00:14:03.942 15:19:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:03.942 15:19:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:03.942 15:19:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:03.942 15:19:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:03.942 15:19:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:14:03.942 15:19:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@652 -- # local es=0 00:14:03.942 15:19:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:14:03.942 15:19:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:14:03.942 15:19:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:14:03.942 15:19:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:14:03.942 15:19:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:14:03.942 15:19:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:14:03.942 15:19:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:03.942 15:19:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:03.942 [2024-11-19 15:19:54.186224] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:03.942 [2024-11-19 15:19:54.186411] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:14:03.942 [2024-11-19 15:19:54.186428] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:14:03.942 request: 00:14:03.942 { 00:14:03.942 "base_bdev": "BaseBdev1", 00:14:03.942 "raid_bdev": "raid_bdev1", 00:14:03.942 "method": "bdev_raid_add_base_bdev", 00:14:03.942 "req_id": 1 00:14:03.942 } 00:14:03.942 Got JSON-RPC error response 00:14:03.942 response: 00:14:03.942 { 00:14:03.942 "code": -22, 00:14:03.942 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:14:03.942 } 00:14:03.942 15:19:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:14:03.942 15:19:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@655 -- # es=1 00:14:03.942 15:19:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:14:03.942 15:19:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:14:03.942 15:19:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:14:03.942 15:19:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:14:04.883 15:19:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:14:04.883 15:19:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:04.883 15:19:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:04.883 15:19:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:04.883 15:19:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:04.883 15:19:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:04.883 15:19:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:04.883 15:19:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:04.883 15:19:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:04.883 15:19:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:04.883 15:19:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:04.883 15:19:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:04.883 15:19:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:04.883 15:19:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:05.143 15:19:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:05.143 15:19:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:05.143 "name": "raid_bdev1", 00:14:05.143 "uuid": "499db5d6-2ee5-45f5-b8aa-305d1e2ff852", 00:14:05.143 "strip_size_kb": 0, 00:14:05.143 "state": "online", 00:14:05.143 "raid_level": "raid1", 00:14:05.143 "superblock": true, 00:14:05.143 "num_base_bdevs": 4, 00:14:05.143 "num_base_bdevs_discovered": 2, 00:14:05.143 "num_base_bdevs_operational": 2, 00:14:05.143 "base_bdevs_list": [ 00:14:05.143 { 00:14:05.143 "name": null, 00:14:05.143 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:05.143 "is_configured": false, 00:14:05.143 "data_offset": 0, 00:14:05.143 "data_size": 63488 00:14:05.143 }, 00:14:05.143 { 00:14:05.143 "name": null, 00:14:05.143 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:05.143 "is_configured": false, 00:14:05.143 "data_offset": 2048, 00:14:05.143 "data_size": 63488 00:14:05.143 }, 00:14:05.143 { 00:14:05.143 "name": "BaseBdev3", 00:14:05.143 "uuid": "bba59b1b-a5ac-5586-a17a-fa7c79f7e077", 00:14:05.143 "is_configured": true, 00:14:05.143 "data_offset": 2048, 00:14:05.143 "data_size": 63488 00:14:05.143 }, 00:14:05.143 { 00:14:05.143 "name": "BaseBdev4", 00:14:05.143 "uuid": "bcf97a61-829b-56dd-ae06-5f30dda1bd20", 00:14:05.143 "is_configured": true, 00:14:05.143 "data_offset": 2048, 00:14:05.143 "data_size": 63488 00:14:05.143 } 00:14:05.143 ] 00:14:05.143 }' 00:14:05.143 15:19:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:05.143 15:19:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:05.404 15:19:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:05.404 15:19:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:05.404 15:19:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:05.404 15:19:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:05.404 15:19:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:05.404 15:19:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:05.404 15:19:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:05.404 15:19:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:05.404 15:19:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:05.404 15:19:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:05.404 15:19:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:05.404 "name": "raid_bdev1", 00:14:05.404 "uuid": "499db5d6-2ee5-45f5-b8aa-305d1e2ff852", 00:14:05.404 "strip_size_kb": 0, 00:14:05.404 "state": "online", 00:14:05.404 "raid_level": "raid1", 00:14:05.404 "superblock": true, 00:14:05.404 "num_base_bdevs": 4, 00:14:05.404 "num_base_bdevs_discovered": 2, 00:14:05.404 "num_base_bdevs_operational": 2, 00:14:05.404 "base_bdevs_list": [ 00:14:05.404 { 00:14:05.404 "name": null, 00:14:05.404 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:05.404 "is_configured": false, 00:14:05.404 "data_offset": 0, 00:14:05.404 "data_size": 63488 00:14:05.404 }, 00:14:05.404 { 00:14:05.404 "name": null, 00:14:05.404 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:05.404 "is_configured": false, 00:14:05.404 "data_offset": 2048, 00:14:05.404 "data_size": 63488 00:14:05.404 }, 00:14:05.404 { 00:14:05.404 "name": "BaseBdev3", 00:14:05.404 "uuid": "bba59b1b-a5ac-5586-a17a-fa7c79f7e077", 00:14:05.404 "is_configured": true, 00:14:05.404 "data_offset": 2048, 00:14:05.404 "data_size": 63488 00:14:05.404 }, 00:14:05.404 { 00:14:05.404 "name": "BaseBdev4", 00:14:05.404 "uuid": "bcf97a61-829b-56dd-ae06-5f30dda1bd20", 00:14:05.404 "is_configured": true, 00:14:05.404 "data_offset": 2048, 00:14:05.404 "data_size": 63488 00:14:05.404 } 00:14:05.404 ] 00:14:05.404 }' 00:14:05.404 15:19:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:05.404 15:19:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:05.404 15:19:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:05.664 15:19:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:05.664 15:19:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 88627 00:14:05.664 15:19:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@954 -- # '[' -z 88627 ']' 00:14:05.664 15:19:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@958 -- # kill -0 88627 00:14:05.664 15:19:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@959 -- # uname 00:14:05.664 15:19:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:14:05.664 15:19:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 88627 00:14:05.664 15:19:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:14:05.664 15:19:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:14:05.664 killing process with pid 88627 00:14:05.664 15:19:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 88627' 00:14:05.664 15:19:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@973 -- # kill 88627 00:14:05.664 Received shutdown signal, test time was about 60.000000 seconds 00:14:05.664 00:14:05.664 Latency(us) 00:14:05.664 [2024-11-19T15:19:56.003Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:14:05.664 [2024-11-19T15:19:56.003Z] =================================================================================================================== 00:14:05.664 [2024-11-19T15:19:56.003Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:14:05.664 [2024-11-19 15:19:55.809254] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:05.664 [2024-11-19 15:19:55.809388] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:05.664 15:19:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@978 -- # wait 88627 00:14:05.664 [2024-11-19 15:19:55.809467] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:05.664 [2024-11-19 15:19:55.809486] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:14:05.664 [2024-11-19 15:19:55.861338] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:05.925 15:19:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:14:05.925 00:14:05.925 real 0m23.572s 00:14:05.925 user 0m28.251s 00:14:05.925 sys 0m4.087s 00:14:05.925 15:19:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:14:05.925 15:19:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:05.925 ************************************ 00:14:05.925 END TEST raid_rebuild_test_sb 00:14:05.925 ************************************ 00:14:05.925 15:19:56 bdev_raid -- bdev/bdev_raid.sh@980 -- # run_test raid_rebuild_test_io raid_rebuild_test raid1 4 false true true 00:14:05.925 15:19:56 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:14:05.925 15:19:56 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:14:05.925 15:19:56 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:05.925 ************************************ 00:14:05.925 START TEST raid_rebuild_test_io 00:14:05.925 ************************************ 00:14:05.925 15:19:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 4 false true true 00:14:05.925 15:19:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:14:05.925 15:19:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:14:05.925 15:19:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:14:05.925 15:19:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:14:05.925 15:19:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:14:05.925 15:19:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:14:05.925 15:19:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:05.925 15:19:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:14:05.925 15:19:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:05.925 15:19:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:05.925 15:19:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:14:05.925 15:19:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:05.925 15:19:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:05.925 15:19:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:14:05.925 15:19:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:05.925 15:19:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:05.925 15:19:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:14:05.925 15:19:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:05.925 15:19:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:05.925 15:19:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:14:05.925 15:19:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:14:05.925 15:19:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:14:05.925 15:19:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:14:05.925 15:19:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:14:05.925 15:19:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:14:05.925 15:19:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:14:05.925 15:19:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:14:05.925 15:19:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:14:05.925 15:19:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:14:05.925 15:19:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@597 -- # raid_pid=89368 00:14:05.925 15:19:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:14:05.925 15:19:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 89368 00:14:05.925 15:19:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@835 -- # '[' -z 89368 ']' 00:14:05.925 15:19:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:05.925 15:19:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@840 -- # local max_retries=100 00:14:05.925 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:05.925 15:19:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:05.925 15:19:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@844 -- # xtrace_disable 00:14:05.925 15:19:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:05.925 I/O size of 3145728 is greater than zero copy threshold (65536). 00:14:05.925 Zero copy mechanism will not be used. 00:14:05.925 [2024-11-19 15:19:56.230049] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:14:05.925 [2024-11-19 15:19:56.230197] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid89368 ] 00:14:06.185 [2024-11-19 15:19:56.363837] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:06.185 [2024-11-19 15:19:56.388902] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:14:06.185 [2024-11-19 15:19:56.431956] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:06.185 [2024-11-19 15:19:56.432022] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:06.755 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:14:06.755 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@868 -- # return 0 00:14:06.755 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:06.755 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:14:06.755 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:06.755 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:06.755 BaseBdev1_malloc 00:14:06.755 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:06.755 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:14:06.755 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:06.755 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:06.755 [2024-11-19 15:19:57.070917] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:14:06.755 [2024-11-19 15:19:57.071004] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:06.755 [2024-11-19 15:19:57.071051] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:14:06.755 [2024-11-19 15:19:57.071075] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:06.755 [2024-11-19 15:19:57.073234] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:06.755 [2024-11-19 15:19:57.073275] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:14:06.755 BaseBdev1 00:14:06.755 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:06.755 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:06.755 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:14:06.755 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:06.755 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:06.756 BaseBdev2_malloc 00:14:06.756 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:06.756 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:14:06.756 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:06.756 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:07.016 [2024-11-19 15:19:57.095624] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:14:07.016 [2024-11-19 15:19:57.095696] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:07.016 [2024-11-19 15:19:57.095728] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:14:07.016 [2024-11-19 15:19:57.095739] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:07.016 [2024-11-19 15:19:57.097824] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:07.016 [2024-11-19 15:19:57.097868] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:14:07.016 BaseBdev2 00:14:07.016 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:07.016 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:07.016 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:14:07.016 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:07.016 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:07.016 BaseBdev3_malloc 00:14:07.016 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:07.016 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:14:07.016 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:07.016 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:07.016 [2024-11-19 15:19:57.124282] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:14:07.016 [2024-11-19 15:19:57.124359] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:07.016 [2024-11-19 15:19:57.124384] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:14:07.016 [2024-11-19 15:19:57.124395] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:07.016 [2024-11-19 15:19:57.126489] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:07.016 [2024-11-19 15:19:57.126526] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:14:07.016 BaseBdev3 00:14:07.016 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:07.017 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:07.017 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:14:07.017 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:07.017 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:07.017 BaseBdev4_malloc 00:14:07.017 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:07.017 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:14:07.017 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:07.017 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:07.017 [2024-11-19 15:19:57.170284] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:14:07.017 [2024-11-19 15:19:57.170498] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:07.017 [2024-11-19 15:19:57.170568] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:14:07.017 [2024-11-19 15:19:57.170596] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:07.017 [2024-11-19 15:19:57.174383] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:07.017 [2024-11-19 15:19:57.174442] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:14:07.017 BaseBdev4 00:14:07.017 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:07.017 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:14:07.017 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:07.017 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:07.017 spare_malloc 00:14:07.017 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:07.017 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:14:07.017 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:07.017 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:07.017 spare_delay 00:14:07.017 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:07.017 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:14:07.017 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:07.017 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:07.017 [2024-11-19 15:19:57.212021] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:14:07.017 [2024-11-19 15:19:57.212067] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:07.017 [2024-11-19 15:19:57.212104] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:14:07.017 [2024-11-19 15:19:57.212115] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:07.017 [2024-11-19 15:19:57.214212] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:07.017 [2024-11-19 15:19:57.214251] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:14:07.017 spare 00:14:07.017 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:07.017 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:14:07.017 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:07.017 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:07.017 [2024-11-19 15:19:57.224074] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:07.017 [2024-11-19 15:19:57.225962] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:07.017 [2024-11-19 15:19:57.226044] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:07.017 [2024-11-19 15:19:57.226095] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:14:07.017 [2024-11-19 15:19:57.226175] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:14:07.017 [2024-11-19 15:19:57.226184] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:14:07.017 [2024-11-19 15:19:57.226419] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:14:07.017 [2024-11-19 15:19:57.226569] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:14:07.017 [2024-11-19 15:19:57.226582] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:14:07.017 [2024-11-19 15:19:57.226716] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:07.017 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:07.017 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:14:07.017 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:07.017 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:07.017 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:07.017 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:07.017 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:07.017 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:07.017 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:07.017 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:07.017 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:07.017 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:07.017 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:07.017 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:07.017 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:07.017 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:07.017 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:07.017 "name": "raid_bdev1", 00:14:07.017 "uuid": "08d210f6-1a8e-4ee9-ba78-df305a1fb857", 00:14:07.017 "strip_size_kb": 0, 00:14:07.017 "state": "online", 00:14:07.017 "raid_level": "raid1", 00:14:07.017 "superblock": false, 00:14:07.017 "num_base_bdevs": 4, 00:14:07.017 "num_base_bdevs_discovered": 4, 00:14:07.017 "num_base_bdevs_operational": 4, 00:14:07.017 "base_bdevs_list": [ 00:14:07.017 { 00:14:07.017 "name": "BaseBdev1", 00:14:07.017 "uuid": "e687b00b-2e10-544d-8e69-7d6a44e8e1bd", 00:14:07.017 "is_configured": true, 00:14:07.017 "data_offset": 0, 00:14:07.017 "data_size": 65536 00:14:07.017 }, 00:14:07.017 { 00:14:07.017 "name": "BaseBdev2", 00:14:07.017 "uuid": "195ccb15-4a88-523d-962d-5d1e574574db", 00:14:07.017 "is_configured": true, 00:14:07.017 "data_offset": 0, 00:14:07.017 "data_size": 65536 00:14:07.017 }, 00:14:07.017 { 00:14:07.017 "name": "BaseBdev3", 00:14:07.017 "uuid": "c3b0cc45-a0e6-552e-b746-fd9cc3fb5dd0", 00:14:07.017 "is_configured": true, 00:14:07.017 "data_offset": 0, 00:14:07.017 "data_size": 65536 00:14:07.017 }, 00:14:07.017 { 00:14:07.017 "name": "BaseBdev4", 00:14:07.017 "uuid": "ab063c1c-4324-5dd3-a102-6a8aaff0242e", 00:14:07.017 "is_configured": true, 00:14:07.017 "data_offset": 0, 00:14:07.017 "data_size": 65536 00:14:07.017 } 00:14:07.017 ] 00:14:07.017 }' 00:14:07.017 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:07.017 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:07.588 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:14:07.588 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:07.588 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:07.588 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:07.588 [2024-11-19 15:19:57.663594] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:07.588 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:07.588 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:14:07.588 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:14:07.588 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:07.588 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:07.588 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:07.588 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:07.588 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:14:07.588 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:14:07.588 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:14:07.588 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:14:07.588 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:07.588 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:07.588 [2024-11-19 15:19:57.731151] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:07.588 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:07.588 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:14:07.588 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:07.588 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:07.588 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:07.588 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:07.588 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:07.588 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:07.588 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:07.588 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:07.588 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:07.588 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:07.588 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:07.588 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:07.588 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:07.588 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:07.588 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:07.588 "name": "raid_bdev1", 00:14:07.588 "uuid": "08d210f6-1a8e-4ee9-ba78-df305a1fb857", 00:14:07.588 "strip_size_kb": 0, 00:14:07.588 "state": "online", 00:14:07.588 "raid_level": "raid1", 00:14:07.588 "superblock": false, 00:14:07.588 "num_base_bdevs": 4, 00:14:07.588 "num_base_bdevs_discovered": 3, 00:14:07.588 "num_base_bdevs_operational": 3, 00:14:07.588 "base_bdevs_list": [ 00:14:07.588 { 00:14:07.588 "name": null, 00:14:07.588 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:07.588 "is_configured": false, 00:14:07.588 "data_offset": 0, 00:14:07.588 "data_size": 65536 00:14:07.588 }, 00:14:07.588 { 00:14:07.588 "name": "BaseBdev2", 00:14:07.588 "uuid": "195ccb15-4a88-523d-962d-5d1e574574db", 00:14:07.588 "is_configured": true, 00:14:07.588 "data_offset": 0, 00:14:07.588 "data_size": 65536 00:14:07.588 }, 00:14:07.588 { 00:14:07.588 "name": "BaseBdev3", 00:14:07.588 "uuid": "c3b0cc45-a0e6-552e-b746-fd9cc3fb5dd0", 00:14:07.588 "is_configured": true, 00:14:07.588 "data_offset": 0, 00:14:07.588 "data_size": 65536 00:14:07.588 }, 00:14:07.588 { 00:14:07.588 "name": "BaseBdev4", 00:14:07.588 "uuid": "ab063c1c-4324-5dd3-a102-6a8aaff0242e", 00:14:07.588 "is_configured": true, 00:14:07.588 "data_offset": 0, 00:14:07.588 "data_size": 65536 00:14:07.588 } 00:14:07.588 ] 00:14:07.588 }' 00:14:07.588 15:19:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:07.588 15:19:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:07.588 [2024-11-19 15:19:57.821129] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:14:07.588 I/O size of 3145728 is greater than zero copy threshold (65536). 00:14:07.588 Zero copy mechanism will not be used. 00:14:07.588 Running I/O for 60 seconds... 00:14:07.848 15:19:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:07.848 15:19:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:07.848 15:19:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:08.108 [2024-11-19 15:19:58.189729] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:08.109 15:19:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:08.109 15:19:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:14:08.109 [2024-11-19 15:19:58.213342] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002940 00:14:08.109 [2024-11-19 15:19:58.215423] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:08.109 [2024-11-19 15:19:58.334814] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:14:08.109 [2024-11-19 15:19:58.336156] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:14:08.368 [2024-11-19 15:19:58.550851] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:14:08.368 [2024-11-19 15:19:58.551280] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:14:08.628 [2024-11-19 15:19:58.816437] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:14:08.628 [2024-11-19 15:19:58.816930] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:14:08.628 178.00 IOPS, 534.00 MiB/s [2024-11-19T15:19:58.967Z] [2024-11-19 15:19:58.926933] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:14:08.628 [2024-11-19 15:19:58.927302] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:14:08.889 [2024-11-19 15:19:59.141777] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:14:08.889 [2024-11-19 15:19:59.142233] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:14:08.889 15:19:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:08.889 15:19:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:08.889 15:19:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:08.889 15:19:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:08.889 15:19:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:08.889 15:19:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:08.889 15:19:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:08.889 15:19:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:08.889 15:19:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:09.148 15:19:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:09.148 15:19:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:09.148 "name": "raid_bdev1", 00:14:09.148 "uuid": "08d210f6-1a8e-4ee9-ba78-df305a1fb857", 00:14:09.148 "strip_size_kb": 0, 00:14:09.148 "state": "online", 00:14:09.148 "raid_level": "raid1", 00:14:09.148 "superblock": false, 00:14:09.148 "num_base_bdevs": 4, 00:14:09.148 "num_base_bdevs_discovered": 4, 00:14:09.148 "num_base_bdevs_operational": 4, 00:14:09.148 "process": { 00:14:09.148 "type": "rebuild", 00:14:09.148 "target": "spare", 00:14:09.148 "progress": { 00:14:09.148 "blocks": 14336, 00:14:09.148 "percent": 21 00:14:09.148 } 00:14:09.148 }, 00:14:09.148 "base_bdevs_list": [ 00:14:09.148 { 00:14:09.148 "name": "spare", 00:14:09.148 "uuid": "4f602980-415b-5a9f-bdf5-be3ea29d6d9e", 00:14:09.148 "is_configured": true, 00:14:09.148 "data_offset": 0, 00:14:09.148 "data_size": 65536 00:14:09.148 }, 00:14:09.148 { 00:14:09.148 "name": "BaseBdev2", 00:14:09.148 "uuid": "195ccb15-4a88-523d-962d-5d1e574574db", 00:14:09.148 "is_configured": true, 00:14:09.148 "data_offset": 0, 00:14:09.148 "data_size": 65536 00:14:09.148 }, 00:14:09.148 { 00:14:09.148 "name": "BaseBdev3", 00:14:09.148 "uuid": "c3b0cc45-a0e6-552e-b746-fd9cc3fb5dd0", 00:14:09.148 "is_configured": true, 00:14:09.148 "data_offset": 0, 00:14:09.148 "data_size": 65536 00:14:09.148 }, 00:14:09.148 { 00:14:09.148 "name": "BaseBdev4", 00:14:09.148 "uuid": "ab063c1c-4324-5dd3-a102-6a8aaff0242e", 00:14:09.148 "is_configured": true, 00:14:09.148 "data_offset": 0, 00:14:09.148 "data_size": 65536 00:14:09.148 } 00:14:09.148 ] 00:14:09.148 }' 00:14:09.148 [2024-11-19 15:19:59.267350] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:14:09.148 15:19:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:09.148 15:19:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:09.148 15:19:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:09.148 15:19:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:09.148 15:19:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:14:09.148 15:19:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:09.148 15:19:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:09.148 [2024-11-19 15:19:59.373533] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:09.408 [2024-11-19 15:19:59.504563] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:14:09.408 [2024-11-19 15:19:59.512274] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:09.408 [2024-11-19 15:19:59.512316] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:09.408 [2024-11-19 15:19:59.512331] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:14:09.408 [2024-11-19 15:19:59.518665] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000002870 00:14:09.408 15:19:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:09.408 15:19:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:14:09.408 15:19:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:09.408 15:19:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:09.408 15:19:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:09.408 15:19:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:09.408 15:19:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:09.408 15:19:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:09.408 15:19:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:09.408 15:19:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:09.408 15:19:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:09.408 15:19:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:09.408 15:19:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:09.408 15:19:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:09.408 15:19:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:09.408 15:19:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:09.408 15:19:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:09.408 "name": "raid_bdev1", 00:14:09.408 "uuid": "08d210f6-1a8e-4ee9-ba78-df305a1fb857", 00:14:09.408 "strip_size_kb": 0, 00:14:09.408 "state": "online", 00:14:09.408 "raid_level": "raid1", 00:14:09.408 "superblock": false, 00:14:09.408 "num_base_bdevs": 4, 00:14:09.408 "num_base_bdevs_discovered": 3, 00:14:09.408 "num_base_bdevs_operational": 3, 00:14:09.408 "base_bdevs_list": [ 00:14:09.408 { 00:14:09.408 "name": null, 00:14:09.408 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:09.408 "is_configured": false, 00:14:09.408 "data_offset": 0, 00:14:09.408 "data_size": 65536 00:14:09.408 }, 00:14:09.408 { 00:14:09.408 "name": "BaseBdev2", 00:14:09.408 "uuid": "195ccb15-4a88-523d-962d-5d1e574574db", 00:14:09.408 "is_configured": true, 00:14:09.408 "data_offset": 0, 00:14:09.408 "data_size": 65536 00:14:09.408 }, 00:14:09.408 { 00:14:09.408 "name": "BaseBdev3", 00:14:09.408 "uuid": "c3b0cc45-a0e6-552e-b746-fd9cc3fb5dd0", 00:14:09.408 "is_configured": true, 00:14:09.408 "data_offset": 0, 00:14:09.408 "data_size": 65536 00:14:09.408 }, 00:14:09.408 { 00:14:09.408 "name": "BaseBdev4", 00:14:09.408 "uuid": "ab063c1c-4324-5dd3-a102-6a8aaff0242e", 00:14:09.408 "is_configured": true, 00:14:09.408 "data_offset": 0, 00:14:09.408 "data_size": 65536 00:14:09.408 } 00:14:09.408 ] 00:14:09.408 }' 00:14:09.408 15:19:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:09.408 15:19:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:09.765 166.50 IOPS, 499.50 MiB/s [2024-11-19T15:20:00.104Z] 15:19:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:09.765 15:19:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:09.765 15:19:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:09.765 15:19:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:09.765 15:19:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:09.765 15:19:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:09.765 15:19:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:09.765 15:19:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:09.765 15:19:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:09.765 15:19:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:09.765 15:19:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:09.765 "name": "raid_bdev1", 00:14:09.765 "uuid": "08d210f6-1a8e-4ee9-ba78-df305a1fb857", 00:14:09.765 "strip_size_kb": 0, 00:14:09.765 "state": "online", 00:14:09.765 "raid_level": "raid1", 00:14:09.765 "superblock": false, 00:14:09.765 "num_base_bdevs": 4, 00:14:09.765 "num_base_bdevs_discovered": 3, 00:14:09.765 "num_base_bdevs_operational": 3, 00:14:09.765 "base_bdevs_list": [ 00:14:09.765 { 00:14:09.765 "name": null, 00:14:09.765 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:09.765 "is_configured": false, 00:14:09.765 "data_offset": 0, 00:14:09.765 "data_size": 65536 00:14:09.765 }, 00:14:09.765 { 00:14:09.765 "name": "BaseBdev2", 00:14:09.765 "uuid": "195ccb15-4a88-523d-962d-5d1e574574db", 00:14:09.765 "is_configured": true, 00:14:09.765 "data_offset": 0, 00:14:09.765 "data_size": 65536 00:14:09.765 }, 00:14:09.765 { 00:14:09.765 "name": "BaseBdev3", 00:14:09.765 "uuid": "c3b0cc45-a0e6-552e-b746-fd9cc3fb5dd0", 00:14:09.765 "is_configured": true, 00:14:09.765 "data_offset": 0, 00:14:09.765 "data_size": 65536 00:14:09.765 }, 00:14:09.765 { 00:14:09.765 "name": "BaseBdev4", 00:14:09.765 "uuid": "ab063c1c-4324-5dd3-a102-6a8aaff0242e", 00:14:09.765 "is_configured": true, 00:14:09.765 "data_offset": 0, 00:14:09.765 "data_size": 65536 00:14:09.765 } 00:14:09.765 ] 00:14:09.765 }' 00:14:09.765 15:19:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:09.765 15:20:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:09.765 15:20:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:10.038 15:20:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:10.038 15:20:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:10.038 15:20:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:10.038 15:20:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:10.038 [2024-11-19 15:20:00.101314] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:10.038 15:20:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:10.038 15:20:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:14:10.038 [2024-11-19 15:20:00.131658] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:14:10.038 [2024-11-19 15:20:00.133664] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:10.038 [2024-11-19 15:20:00.247409] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:14:10.038 [2024-11-19 15:20:00.248593] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:14:10.298 [2024-11-19 15:20:00.464160] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:14:10.298 [2024-11-19 15:20:00.464804] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:14:10.818 158.67 IOPS, 476.00 MiB/s [2024-11-19T15:20:01.157Z] [2024-11-19 15:20:00.944714] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:14:10.818 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:10.818 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:10.818 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:10.818 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:10.818 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:10.818 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:10.818 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:10.818 15:20:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:10.818 15:20:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:11.077 15:20:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:11.077 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:11.077 "name": "raid_bdev1", 00:14:11.077 "uuid": "08d210f6-1a8e-4ee9-ba78-df305a1fb857", 00:14:11.077 "strip_size_kb": 0, 00:14:11.077 "state": "online", 00:14:11.077 "raid_level": "raid1", 00:14:11.077 "superblock": false, 00:14:11.077 "num_base_bdevs": 4, 00:14:11.077 "num_base_bdevs_discovered": 4, 00:14:11.077 "num_base_bdevs_operational": 4, 00:14:11.077 "process": { 00:14:11.077 "type": "rebuild", 00:14:11.077 "target": "spare", 00:14:11.077 "progress": { 00:14:11.077 "blocks": 12288, 00:14:11.077 "percent": 18 00:14:11.077 } 00:14:11.077 }, 00:14:11.077 "base_bdevs_list": [ 00:14:11.077 { 00:14:11.077 "name": "spare", 00:14:11.077 "uuid": "4f602980-415b-5a9f-bdf5-be3ea29d6d9e", 00:14:11.077 "is_configured": true, 00:14:11.077 "data_offset": 0, 00:14:11.077 "data_size": 65536 00:14:11.077 }, 00:14:11.077 { 00:14:11.077 "name": "BaseBdev2", 00:14:11.077 "uuid": "195ccb15-4a88-523d-962d-5d1e574574db", 00:14:11.077 "is_configured": true, 00:14:11.077 "data_offset": 0, 00:14:11.077 "data_size": 65536 00:14:11.077 }, 00:14:11.077 { 00:14:11.077 "name": "BaseBdev3", 00:14:11.078 "uuid": "c3b0cc45-a0e6-552e-b746-fd9cc3fb5dd0", 00:14:11.078 "is_configured": true, 00:14:11.078 "data_offset": 0, 00:14:11.078 "data_size": 65536 00:14:11.078 }, 00:14:11.078 { 00:14:11.078 "name": "BaseBdev4", 00:14:11.078 "uuid": "ab063c1c-4324-5dd3-a102-6a8aaff0242e", 00:14:11.078 "is_configured": true, 00:14:11.078 "data_offset": 0, 00:14:11.078 "data_size": 65536 00:14:11.078 } 00:14:11.078 ] 00:14:11.078 }' 00:14:11.078 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:11.078 [2024-11-19 15:20:01.186902] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:14:11.078 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:11.078 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:11.078 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:11.078 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:14:11.078 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:14:11.078 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:14:11.078 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:14:11.078 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:14:11.078 15:20:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:11.078 15:20:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:11.078 [2024-11-19 15:20:01.284791] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:14:11.078 [2024-11-19 15:20:01.411145] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:14:11.338 [2024-11-19 15:20:01.513426] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000002870 00:14:11.338 [2024-11-19 15:20:01.513472] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000002a10 00:14:11.338 15:20:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:11.338 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:14:11.338 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:14:11.338 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:11.338 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:11.338 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:11.338 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:11.338 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:11.338 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:11.338 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:11.338 15:20:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:11.338 15:20:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:11.338 15:20:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:11.338 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:11.338 "name": "raid_bdev1", 00:14:11.338 "uuid": "08d210f6-1a8e-4ee9-ba78-df305a1fb857", 00:14:11.338 "strip_size_kb": 0, 00:14:11.338 "state": "online", 00:14:11.338 "raid_level": "raid1", 00:14:11.338 "superblock": false, 00:14:11.338 "num_base_bdevs": 4, 00:14:11.338 "num_base_bdevs_discovered": 3, 00:14:11.338 "num_base_bdevs_operational": 3, 00:14:11.338 "process": { 00:14:11.338 "type": "rebuild", 00:14:11.338 "target": "spare", 00:14:11.338 "progress": { 00:14:11.338 "blocks": 16384, 00:14:11.338 "percent": 25 00:14:11.338 } 00:14:11.338 }, 00:14:11.338 "base_bdevs_list": [ 00:14:11.338 { 00:14:11.338 "name": "spare", 00:14:11.338 "uuid": "4f602980-415b-5a9f-bdf5-be3ea29d6d9e", 00:14:11.338 "is_configured": true, 00:14:11.338 "data_offset": 0, 00:14:11.338 "data_size": 65536 00:14:11.338 }, 00:14:11.338 { 00:14:11.338 "name": null, 00:14:11.338 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:11.338 "is_configured": false, 00:14:11.338 "data_offset": 0, 00:14:11.338 "data_size": 65536 00:14:11.338 }, 00:14:11.338 { 00:14:11.338 "name": "BaseBdev3", 00:14:11.338 "uuid": "c3b0cc45-a0e6-552e-b746-fd9cc3fb5dd0", 00:14:11.338 "is_configured": true, 00:14:11.338 "data_offset": 0, 00:14:11.338 "data_size": 65536 00:14:11.338 }, 00:14:11.338 { 00:14:11.338 "name": "BaseBdev4", 00:14:11.338 "uuid": "ab063c1c-4324-5dd3-a102-6a8aaff0242e", 00:14:11.338 "is_configured": true, 00:14:11.338 "data_offset": 0, 00:14:11.338 "data_size": 65536 00:14:11.338 } 00:14:11.338 ] 00:14:11.338 }' 00:14:11.338 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:11.338 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:11.338 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:11.338 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:11.338 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@706 -- # local timeout=399 00:14:11.338 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:11.338 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:11.338 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:11.338 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:11.338 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:11.338 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:11.339 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:11.339 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:11.339 15:20:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:11.339 15:20:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:11.599 15:20:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:11.599 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:11.600 "name": "raid_bdev1", 00:14:11.600 "uuid": "08d210f6-1a8e-4ee9-ba78-df305a1fb857", 00:14:11.600 "strip_size_kb": 0, 00:14:11.600 "state": "online", 00:14:11.600 "raid_level": "raid1", 00:14:11.600 "superblock": false, 00:14:11.600 "num_base_bdevs": 4, 00:14:11.600 "num_base_bdevs_discovered": 3, 00:14:11.600 "num_base_bdevs_operational": 3, 00:14:11.600 "process": { 00:14:11.600 "type": "rebuild", 00:14:11.600 "target": "spare", 00:14:11.600 "progress": { 00:14:11.600 "blocks": 18432, 00:14:11.600 "percent": 28 00:14:11.600 } 00:14:11.600 }, 00:14:11.600 "base_bdevs_list": [ 00:14:11.600 { 00:14:11.600 "name": "spare", 00:14:11.600 "uuid": "4f602980-415b-5a9f-bdf5-be3ea29d6d9e", 00:14:11.600 "is_configured": true, 00:14:11.600 "data_offset": 0, 00:14:11.600 "data_size": 65536 00:14:11.600 }, 00:14:11.600 { 00:14:11.600 "name": null, 00:14:11.600 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:11.600 "is_configured": false, 00:14:11.600 "data_offset": 0, 00:14:11.600 "data_size": 65536 00:14:11.600 }, 00:14:11.600 { 00:14:11.600 "name": "BaseBdev3", 00:14:11.600 "uuid": "c3b0cc45-a0e6-552e-b746-fd9cc3fb5dd0", 00:14:11.600 "is_configured": true, 00:14:11.600 "data_offset": 0, 00:14:11.600 "data_size": 65536 00:14:11.600 }, 00:14:11.600 { 00:14:11.600 "name": "BaseBdev4", 00:14:11.600 "uuid": "ab063c1c-4324-5dd3-a102-6a8aaff0242e", 00:14:11.600 "is_configured": true, 00:14:11.600 "data_offset": 0, 00:14:11.600 "data_size": 65536 00:14:11.600 } 00:14:11.600 ] 00:14:11.600 }' 00:14:11.600 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:11.600 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:11.600 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:11.600 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:11.600 15:20:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:11.600 [2024-11-19 15:20:01.754136] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:14:11.600 [2024-11-19 15:20:01.754943] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:14:11.860 144.50 IOPS, 433.50 MiB/s [2024-11-19T15:20:02.199Z] [2024-11-19 15:20:01.969640] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:14:12.120 [2024-11-19 15:20:02.310847] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 26624 offset_begin: 24576 offset_end: 30720 00:14:12.380 [2024-11-19 15:20:02.524241] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 28672 offset_begin: 24576 offset_end: 30720 00:14:12.647 15:20:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:12.647 15:20:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:12.647 15:20:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:12.647 15:20:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:12.647 15:20:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:12.647 15:20:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:12.647 15:20:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:12.647 15:20:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:12.647 15:20:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:12.647 15:20:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:12.647 15:20:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:12.647 15:20:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:12.647 "name": "raid_bdev1", 00:14:12.647 "uuid": "08d210f6-1a8e-4ee9-ba78-df305a1fb857", 00:14:12.647 "strip_size_kb": 0, 00:14:12.647 "state": "online", 00:14:12.647 "raid_level": "raid1", 00:14:12.647 "superblock": false, 00:14:12.647 "num_base_bdevs": 4, 00:14:12.647 "num_base_bdevs_discovered": 3, 00:14:12.647 "num_base_bdevs_operational": 3, 00:14:12.647 "process": { 00:14:12.647 "type": "rebuild", 00:14:12.647 "target": "spare", 00:14:12.647 "progress": { 00:14:12.647 "blocks": 30720, 00:14:12.647 "percent": 46 00:14:12.647 } 00:14:12.647 }, 00:14:12.647 "base_bdevs_list": [ 00:14:12.647 { 00:14:12.647 "name": "spare", 00:14:12.647 "uuid": "4f602980-415b-5a9f-bdf5-be3ea29d6d9e", 00:14:12.647 "is_configured": true, 00:14:12.647 "data_offset": 0, 00:14:12.647 "data_size": 65536 00:14:12.647 }, 00:14:12.647 { 00:14:12.647 "name": null, 00:14:12.647 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:12.647 "is_configured": false, 00:14:12.647 "data_offset": 0, 00:14:12.647 "data_size": 65536 00:14:12.647 }, 00:14:12.647 { 00:14:12.647 "name": "BaseBdev3", 00:14:12.647 "uuid": "c3b0cc45-a0e6-552e-b746-fd9cc3fb5dd0", 00:14:12.647 "is_configured": true, 00:14:12.647 "data_offset": 0, 00:14:12.647 "data_size": 65536 00:14:12.647 }, 00:14:12.647 { 00:14:12.647 "name": "BaseBdev4", 00:14:12.647 "uuid": "ab063c1c-4324-5dd3-a102-6a8aaff0242e", 00:14:12.647 "is_configured": true, 00:14:12.647 "data_offset": 0, 00:14:12.647 "data_size": 65536 00:14:12.647 } 00:14:12.647 ] 00:14:12.647 }' 00:14:12.647 15:20:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:12.647 125.60 IOPS, 376.80 MiB/s [2024-11-19T15:20:02.986Z] 15:20:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:12.647 15:20:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:12.647 15:20:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:12.647 15:20:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:12.647 [2024-11-19 15:20:02.956444] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 34816 offset_begin: 30720 offset_end: 36864 00:14:13.217 [2024-11-19 15:20:03.420784] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 40960 offset_begin: 36864 offset_end: 43008 00:14:13.476 [2024-11-19 15:20:03.746047] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 45056 offset_begin: 43008 offset_end: 49152 00:14:13.476 [2024-11-19 15:20:03.746925] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 45056 offset_begin: 43008 offset_end: 49152 00:14:13.737 110.00 IOPS, 330.00 MiB/s [2024-11-19T15:20:04.076Z] 15:20:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:13.737 15:20:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:13.737 15:20:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:13.737 15:20:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:13.737 15:20:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:13.737 15:20:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:13.737 15:20:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:13.737 15:20:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:13.737 15:20:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:13.737 15:20:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:13.737 15:20:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:13.737 [2024-11-19 15:20:03.950011] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 47104 offset_begin: 43008 offset_end: 49152 00:14:13.737 15:20:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:13.737 "name": "raid_bdev1", 00:14:13.737 "uuid": "08d210f6-1a8e-4ee9-ba78-df305a1fb857", 00:14:13.737 "strip_size_kb": 0, 00:14:13.737 "state": "online", 00:14:13.737 "raid_level": "raid1", 00:14:13.737 "superblock": false, 00:14:13.737 "num_base_bdevs": 4, 00:14:13.737 "num_base_bdevs_discovered": 3, 00:14:13.737 "num_base_bdevs_operational": 3, 00:14:13.737 "process": { 00:14:13.737 "type": "rebuild", 00:14:13.737 "target": "spare", 00:14:13.737 "progress": { 00:14:13.737 "blocks": 45056, 00:14:13.737 "percent": 68 00:14:13.737 } 00:14:13.737 }, 00:14:13.737 "base_bdevs_list": [ 00:14:13.737 { 00:14:13.737 "name": "spare", 00:14:13.737 "uuid": "4f602980-415b-5a9f-bdf5-be3ea29d6d9e", 00:14:13.737 "is_configured": true, 00:14:13.737 "data_offset": 0, 00:14:13.737 "data_size": 65536 00:14:13.737 }, 00:14:13.737 { 00:14:13.737 "name": null, 00:14:13.737 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:13.737 "is_configured": false, 00:14:13.737 "data_offset": 0, 00:14:13.737 "data_size": 65536 00:14:13.737 }, 00:14:13.737 { 00:14:13.737 "name": "BaseBdev3", 00:14:13.737 "uuid": "c3b0cc45-a0e6-552e-b746-fd9cc3fb5dd0", 00:14:13.737 "is_configured": true, 00:14:13.737 "data_offset": 0, 00:14:13.737 "data_size": 65536 00:14:13.737 }, 00:14:13.737 { 00:14:13.737 "name": "BaseBdev4", 00:14:13.737 "uuid": "ab063c1c-4324-5dd3-a102-6a8aaff0242e", 00:14:13.737 "is_configured": true, 00:14:13.737 "data_offset": 0, 00:14:13.737 "data_size": 65536 00:14:13.737 } 00:14:13.737 ] 00:14:13.737 }' 00:14:13.737 15:20:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:13.737 15:20:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:13.737 15:20:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:13.737 15:20:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:13.737 15:20:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:14.677 98.71 IOPS, 296.14 MiB/s [2024-11-19T15:20:05.016Z] [2024-11-19 15:20:04.917497] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:14:14.937 [2024-11-19 15:20:05.017368] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:14:14.937 [2024-11-19 15:20:05.019549] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:14.937 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:14.937 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:14.937 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:14.937 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:14.937 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:14.937 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:14.937 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:14.937 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:14.937 15:20:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:14.937 15:20:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:14.937 15:20:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:14.937 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:14.937 "name": "raid_bdev1", 00:14:14.937 "uuid": "08d210f6-1a8e-4ee9-ba78-df305a1fb857", 00:14:14.937 "strip_size_kb": 0, 00:14:14.937 "state": "online", 00:14:14.937 "raid_level": "raid1", 00:14:14.937 "superblock": false, 00:14:14.937 "num_base_bdevs": 4, 00:14:14.937 "num_base_bdevs_discovered": 3, 00:14:14.937 "num_base_bdevs_operational": 3, 00:14:14.937 "base_bdevs_list": [ 00:14:14.937 { 00:14:14.937 "name": "spare", 00:14:14.937 "uuid": "4f602980-415b-5a9f-bdf5-be3ea29d6d9e", 00:14:14.937 "is_configured": true, 00:14:14.937 "data_offset": 0, 00:14:14.937 "data_size": 65536 00:14:14.937 }, 00:14:14.937 { 00:14:14.937 "name": null, 00:14:14.937 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:14.937 "is_configured": false, 00:14:14.937 "data_offset": 0, 00:14:14.937 "data_size": 65536 00:14:14.937 }, 00:14:14.937 { 00:14:14.937 "name": "BaseBdev3", 00:14:14.937 "uuid": "c3b0cc45-a0e6-552e-b746-fd9cc3fb5dd0", 00:14:14.937 "is_configured": true, 00:14:14.937 "data_offset": 0, 00:14:14.937 "data_size": 65536 00:14:14.937 }, 00:14:14.937 { 00:14:14.938 "name": "BaseBdev4", 00:14:14.938 "uuid": "ab063c1c-4324-5dd3-a102-6a8aaff0242e", 00:14:14.938 "is_configured": true, 00:14:14.938 "data_offset": 0, 00:14:14.938 "data_size": 65536 00:14:14.938 } 00:14:14.938 ] 00:14:14.938 }' 00:14:14.938 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:14.938 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:14:14.938 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:14.938 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:14:14.938 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@709 -- # break 00:14:14.938 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:14.938 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:14.938 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:14.938 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:14.938 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:14.938 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:14.938 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:14.938 15:20:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:14.938 15:20:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:14.938 15:20:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:14.938 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:14.938 "name": "raid_bdev1", 00:14:14.938 "uuid": "08d210f6-1a8e-4ee9-ba78-df305a1fb857", 00:14:14.938 "strip_size_kb": 0, 00:14:14.938 "state": "online", 00:14:14.938 "raid_level": "raid1", 00:14:14.938 "superblock": false, 00:14:14.938 "num_base_bdevs": 4, 00:14:14.938 "num_base_bdevs_discovered": 3, 00:14:14.938 "num_base_bdevs_operational": 3, 00:14:14.938 "base_bdevs_list": [ 00:14:14.938 { 00:14:14.938 "name": "spare", 00:14:14.938 "uuid": "4f602980-415b-5a9f-bdf5-be3ea29d6d9e", 00:14:14.938 "is_configured": true, 00:14:14.938 "data_offset": 0, 00:14:14.938 "data_size": 65536 00:14:14.938 }, 00:14:14.938 { 00:14:14.938 "name": null, 00:14:14.938 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:14.938 "is_configured": false, 00:14:14.938 "data_offset": 0, 00:14:14.938 "data_size": 65536 00:14:14.938 }, 00:14:14.938 { 00:14:14.938 "name": "BaseBdev3", 00:14:14.938 "uuid": "c3b0cc45-a0e6-552e-b746-fd9cc3fb5dd0", 00:14:14.938 "is_configured": true, 00:14:14.938 "data_offset": 0, 00:14:14.938 "data_size": 65536 00:14:14.938 }, 00:14:14.938 { 00:14:14.938 "name": "BaseBdev4", 00:14:14.938 "uuid": "ab063c1c-4324-5dd3-a102-6a8aaff0242e", 00:14:14.938 "is_configured": true, 00:14:14.938 "data_offset": 0, 00:14:14.938 "data_size": 65536 00:14:14.938 } 00:14:14.938 ] 00:14:14.938 }' 00:14:14.938 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:14.938 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:14.938 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:15.198 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:15.198 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:14:15.198 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:15.198 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:15.198 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:15.198 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:15.198 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:15.198 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:15.198 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:15.198 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:15.198 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:15.198 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:15.198 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:15.198 15:20:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:15.198 15:20:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:15.198 15:20:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:15.198 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:15.198 "name": "raid_bdev1", 00:14:15.198 "uuid": "08d210f6-1a8e-4ee9-ba78-df305a1fb857", 00:14:15.198 "strip_size_kb": 0, 00:14:15.199 "state": "online", 00:14:15.199 "raid_level": "raid1", 00:14:15.199 "superblock": false, 00:14:15.199 "num_base_bdevs": 4, 00:14:15.199 "num_base_bdevs_discovered": 3, 00:14:15.199 "num_base_bdevs_operational": 3, 00:14:15.199 "base_bdevs_list": [ 00:14:15.199 { 00:14:15.199 "name": "spare", 00:14:15.199 "uuid": "4f602980-415b-5a9f-bdf5-be3ea29d6d9e", 00:14:15.199 "is_configured": true, 00:14:15.199 "data_offset": 0, 00:14:15.199 "data_size": 65536 00:14:15.199 }, 00:14:15.199 { 00:14:15.199 "name": null, 00:14:15.199 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:15.199 "is_configured": false, 00:14:15.199 "data_offset": 0, 00:14:15.199 "data_size": 65536 00:14:15.199 }, 00:14:15.199 { 00:14:15.199 "name": "BaseBdev3", 00:14:15.199 "uuid": "c3b0cc45-a0e6-552e-b746-fd9cc3fb5dd0", 00:14:15.199 "is_configured": true, 00:14:15.199 "data_offset": 0, 00:14:15.199 "data_size": 65536 00:14:15.199 }, 00:14:15.199 { 00:14:15.199 "name": "BaseBdev4", 00:14:15.199 "uuid": "ab063c1c-4324-5dd3-a102-6a8aaff0242e", 00:14:15.199 "is_configured": true, 00:14:15.199 "data_offset": 0, 00:14:15.199 "data_size": 65536 00:14:15.199 } 00:14:15.199 ] 00:14:15.199 }' 00:14:15.199 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:15.199 15:20:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:15.459 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:15.459 15:20:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:15.459 15:20:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:15.459 [2024-11-19 15:20:05.720883] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:15.459 [2024-11-19 15:20:05.721006] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:15.459 00:14:15.459 Latency(us) 00:14:15.459 [2024-11-19T15:20:05.798Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:14:15.459 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:14:15.459 raid_bdev1 : 7.96 91.40 274.21 0.00 0.00 13846.72 284.39 116762.83 00:14:15.459 [2024-11-19T15:20:05.798Z] =================================================================================================================== 00:14:15.459 [2024-11-19T15:20:05.798Z] Total : 91.40 274.21 0.00 0.00 13846.72 284.39 116762.83 00:14:15.459 [2024-11-19 15:20:05.775945] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:15.459 [2024-11-19 15:20:05.776048] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:15.459 [2024-11-19 15:20:05.776176] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:15.459 [2024-11-19 15:20:05.776240] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:14:15.459 { 00:14:15.459 "results": [ 00:14:15.459 { 00:14:15.459 "job": "raid_bdev1", 00:14:15.459 "core_mask": "0x1", 00:14:15.459 "workload": "randrw", 00:14:15.459 "percentage": 50, 00:14:15.459 "status": "finished", 00:14:15.459 "queue_depth": 2, 00:14:15.459 "io_size": 3145728, 00:14:15.459 "runtime": 7.964663, 00:14:15.459 "iops": 91.4037417527898, 00:14:15.459 "mibps": 274.2112252583694, 00:14:15.459 "io_failed": 0, 00:14:15.459 "io_timeout": 0, 00:14:15.459 "avg_latency_us": 13846.72416142809, 00:14:15.459 "min_latency_us": 284.3947598253275, 00:14:15.459 "max_latency_us": 116762.82969432314 00:14:15.459 } 00:14:15.459 ], 00:14:15.459 "core_count": 1 00:14:15.459 } 00:14:15.459 15:20:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:15.459 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:15.459 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # jq length 00:14:15.459 15:20:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:15.459 15:20:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:15.459 15:20:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:15.719 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:14:15.720 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:14:15.720 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:14:15.720 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:14:15.720 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:14:15.720 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:14:15.720 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:14:15.720 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:14:15.720 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:14:15.720 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:14:15.720 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:14:15.720 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:15.720 15:20:05 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:14:15.720 /dev/nbd0 00:14:15.720 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:14:15.720 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:14:15.720 15:20:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:14:15.720 15:20:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # local i 00:14:15.720 15:20:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:14:15.720 15:20:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:14:15.720 15:20:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:14:15.980 15:20:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@877 -- # break 00:14:15.980 15:20:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:14:15.980 15:20:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:14:15.980 15:20:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:15.980 1+0 records in 00:14:15.980 1+0 records out 00:14:15.980 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00140558 s, 2.9 MB/s 00:14:15.980 15:20:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:15.980 15:20:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # size=4096 00:14:15.980 15:20:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:15.980 15:20:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:14:15.980 15:20:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@893 -- # return 0 00:14:15.980 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:15.980 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:15.980 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:14:15.980 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z '' ']' 00:14:15.980 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@728 -- # continue 00:14:15.980 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:14:15.980 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev3 ']' 00:14:15.980 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev3 /dev/nbd1 00:14:15.980 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:14:15.980 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev3') 00:14:15.980 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:14:15.980 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:14:15.980 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:14:15.980 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:14:15.980 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:14:15.980 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:15.980 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev3 /dev/nbd1 00:14:15.980 /dev/nbd1 00:14:15.980 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:14:15.980 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:14:15.980 15:20:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:14:15.980 15:20:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # local i 00:14:15.980 15:20:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:14:15.980 15:20:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:14:15.980 15:20:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:14:15.980 15:20:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@877 -- # break 00:14:15.980 15:20:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:14:15.980 15:20:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:14:15.980 15:20:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:15.980 1+0 records in 00:14:15.980 1+0 records out 00:14:15.980 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000381011 s, 10.8 MB/s 00:14:16.241 15:20:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:16.241 15:20:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # size=4096 00:14:16.241 15:20:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:16.241 15:20:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:14:16.241 15:20:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@893 -- # return 0 00:14:16.241 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:16.241 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:16.241 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@731 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:14:16.241 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:14:16.241 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:14:16.241 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:14:16.241 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:14:16.241 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:14:16.241 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:16.241 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:14:16.501 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:14:16.501 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:14:16.501 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:14:16.501 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:16.501 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:16.501 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:14:16.501 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:14:16.501 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:14:16.501 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:14:16.501 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev4 ']' 00:14:16.501 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev4 /dev/nbd1 00:14:16.501 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:14:16.501 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev4') 00:14:16.501 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:14:16.501 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:14:16.501 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:14:16.501 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:14:16.501 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:14:16.501 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:16.501 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev4 /dev/nbd1 00:14:16.762 /dev/nbd1 00:14:16.762 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:14:16.762 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:14:16.762 15:20:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:14:16.762 15:20:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # local i 00:14:16.762 15:20:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:14:16.762 15:20:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:14:16.762 15:20:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:14:16.762 15:20:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@877 -- # break 00:14:16.762 15:20:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:14:16.762 15:20:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:14:16.762 15:20:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:16.762 1+0 records in 00:14:16.762 1+0 records out 00:14:16.762 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000511883 s, 8.0 MB/s 00:14:16.762 15:20:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:16.762 15:20:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # size=4096 00:14:16.762 15:20:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:16.762 15:20:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:14:16.762 15:20:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@893 -- # return 0 00:14:16.762 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:16.762 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:16.762 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@731 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:14:16.762 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:14:16.762 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:14:16.762 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:14:16.762 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:14:16.762 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:14:16.762 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:16.762 15:20:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:14:17.022 15:20:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:14:17.022 15:20:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:14:17.022 15:20:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:14:17.022 15:20:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:17.022 15:20:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:17.022 15:20:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:14:17.022 15:20:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:14:17.022 15:20:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:14:17.022 15:20:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:14:17.022 15:20:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:14:17.022 15:20:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:14:17.022 15:20:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:14:17.022 15:20:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:14:17.022 15:20:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:17.022 15:20:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:14:17.022 15:20:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:14:17.022 15:20:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:14:17.022 15:20:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:14:17.022 15:20:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:17.022 15:20:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:17.022 15:20:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:14:17.282 15:20:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:14:17.282 15:20:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:14:17.282 15:20:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:14:17.282 15:20:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@784 -- # killprocess 89368 00:14:17.282 15:20:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@954 -- # '[' -z 89368 ']' 00:14:17.282 15:20:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@958 -- # kill -0 89368 00:14:17.282 15:20:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@959 -- # uname 00:14:17.282 15:20:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:14:17.282 15:20:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 89368 00:14:17.282 killing process with pid 89368 00:14:17.282 Received shutdown signal, test time was about 9.597521 seconds 00:14:17.282 00:14:17.282 Latency(us) 00:14:17.282 [2024-11-19T15:20:07.621Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:14:17.282 [2024-11-19T15:20:07.621Z] =================================================================================================================== 00:14:17.282 [2024-11-19T15:20:07.621Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:14:17.282 15:20:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:14:17.282 15:20:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:14:17.282 15:20:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@972 -- # echo 'killing process with pid 89368' 00:14:17.282 15:20:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@973 -- # kill 89368 00:14:17.282 [2024-11-19 15:20:07.402421] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:17.282 15:20:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@978 -- # wait 89368 00:14:17.282 [2024-11-19 15:20:07.490263] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:17.542 15:20:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@786 -- # return 0 00:14:17.542 00:14:17.542 real 0m11.668s 00:14:17.542 user 0m14.956s 00:14:17.542 sys 0m1.764s 00:14:17.542 15:20:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1130 -- # xtrace_disable 00:14:17.542 ************************************ 00:14:17.542 END TEST raid_rebuild_test_io 00:14:17.542 ************************************ 00:14:17.542 15:20:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:14:17.542 15:20:07 bdev_raid -- bdev/bdev_raid.sh@981 -- # run_test raid_rebuild_test_sb_io raid_rebuild_test raid1 4 true true true 00:14:17.542 15:20:07 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:14:17.542 15:20:07 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:14:17.542 15:20:07 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:17.802 ************************************ 00:14:17.802 START TEST raid_rebuild_test_sb_io 00:14:17.802 ************************************ 00:14:17.802 15:20:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 4 true true true 00:14:17.802 15:20:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:14:17.802 15:20:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:14:17.802 15:20:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:14:17.802 15:20:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:14:17.802 15:20:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:14:17.802 15:20:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:14:17.802 15:20:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:17.802 15:20:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:14:17.802 15:20:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:17.802 15:20:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:17.802 15:20:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:14:17.802 15:20:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:17.802 15:20:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:17.802 15:20:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:14:17.802 15:20:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:17.802 15:20:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:17.802 15:20:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:14:17.802 15:20:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:17.802 15:20:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:17.802 15:20:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:14:17.802 15:20:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:14:17.802 15:20:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:14:17.802 15:20:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:14:17.802 15:20:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:14:17.802 15:20:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:14:17.802 15:20:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:14:17.802 15:20:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:14:17.802 15:20:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:14:17.802 15:20:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:14:17.802 15:20:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:14:17.802 15:20:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@597 -- # raid_pid=89761 00:14:17.802 15:20:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:14:17.802 15:20:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 89761 00:14:17.802 15:20:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@835 -- # '[' -z 89761 ']' 00:14:17.802 15:20:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:17.802 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:17.802 15:20:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@840 -- # local max_retries=100 00:14:17.802 15:20:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:17.802 15:20:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@844 -- # xtrace_disable 00:14:17.802 15:20:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:17.802 I/O size of 3145728 is greater than zero copy threshold (65536). 00:14:17.802 Zero copy mechanism will not be used. 00:14:17.802 [2024-11-19 15:20:07.981398] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:14:17.802 [2024-11-19 15:20:07.981505] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid89761 ] 00:14:17.802 [2024-11-19 15:20:08.112266] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:18.062 [2024-11-19 15:20:08.151412] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:14:18.062 [2024-11-19 15:20:08.227302] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:18.062 [2024-11-19 15:20:08.227347] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:18.632 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:14:18.632 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@868 -- # return 0 00:14:18.632 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:18.632 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:14:18.632 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:18.632 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:18.632 BaseBdev1_malloc 00:14:18.632 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:18.632 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:14:18.632 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:18.632 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:18.632 [2024-11-19 15:20:08.820497] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:14:18.632 [2024-11-19 15:20:08.820573] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:18.632 [2024-11-19 15:20:08.820604] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:14:18.632 [2024-11-19 15:20:08.820617] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:18.632 [2024-11-19 15:20:08.822959] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:18.632 [2024-11-19 15:20:08.823004] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:14:18.632 BaseBdev1 00:14:18.632 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:18.632 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:18.632 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:14:18.632 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:18.632 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:18.632 BaseBdev2_malloc 00:14:18.632 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:18.632 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:14:18.632 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:18.632 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:18.632 [2024-11-19 15:20:08.854836] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:14:18.632 [2024-11-19 15:20:08.854959] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:18.632 [2024-11-19 15:20:08.854995] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:14:18.632 [2024-11-19 15:20:08.855004] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:18.632 [2024-11-19 15:20:08.857312] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:18.632 [2024-11-19 15:20:08.857350] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:14:18.632 BaseBdev2 00:14:18.632 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:18.632 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:18.632 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:14:18.632 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:18.632 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:18.632 BaseBdev3_malloc 00:14:18.632 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:18.632 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:14:18.632 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:18.632 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:18.632 [2024-11-19 15:20:08.889175] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:14:18.632 [2024-11-19 15:20:08.889297] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:18.632 [2024-11-19 15:20:08.889326] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:14:18.632 [2024-11-19 15:20:08.889335] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:18.632 [2024-11-19 15:20:08.891650] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:18.632 [2024-11-19 15:20:08.891683] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:14:18.632 BaseBdev3 00:14:18.632 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:18.632 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:18.632 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:14:18.633 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:18.633 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:18.633 BaseBdev4_malloc 00:14:18.633 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:18.633 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:14:18.633 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:18.633 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:18.633 [2024-11-19 15:20:08.935026] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:14:18.633 [2024-11-19 15:20:08.935072] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:18.633 [2024-11-19 15:20:08.935096] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:14:18.633 [2024-11-19 15:20:08.935104] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:18.633 [2024-11-19 15:20:08.937455] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:18.633 [2024-11-19 15:20:08.937489] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:14:18.633 BaseBdev4 00:14:18.633 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:18.633 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:14:18.633 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:18.633 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:18.633 spare_malloc 00:14:18.633 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:18.633 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:14:18.633 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:18.633 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:18.893 spare_delay 00:14:18.893 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:18.893 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:14:18.893 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:18.893 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:18.893 [2024-11-19 15:20:08.981493] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:14:18.893 [2024-11-19 15:20:08.981589] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:18.893 [2024-11-19 15:20:08.981612] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:14:18.893 [2024-11-19 15:20:08.981621] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:18.893 [2024-11-19 15:20:08.983882] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:18.893 [2024-11-19 15:20:08.983919] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:14:18.893 spare 00:14:18.893 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:18.893 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:14:18.893 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:18.893 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:18.893 [2024-11-19 15:20:08.993552] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:18.893 [2024-11-19 15:20:08.995510] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:18.893 [2024-11-19 15:20:08.995572] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:18.893 [2024-11-19 15:20:08.995614] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:14:18.893 [2024-11-19 15:20:08.995799] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:14:18.893 [2024-11-19 15:20:08.995819] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:14:18.893 [2024-11-19 15:20:08.996088] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:14:18.893 [2024-11-19 15:20:08.996256] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:14:18.893 [2024-11-19 15:20:08.996269] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:14:18.893 [2024-11-19 15:20:08.996383] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:18.893 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:18.893 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:14:18.893 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:18.893 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:18.893 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:18.893 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:18.893 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:18.893 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:18.893 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:18.893 15:20:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:18.893 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:18.893 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:18.893 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:18.893 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:18.893 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:18.893 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:18.893 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:18.893 "name": "raid_bdev1", 00:14:18.893 "uuid": "a426f079-d827-4102-96e5-ce4cc6c6d4c0", 00:14:18.893 "strip_size_kb": 0, 00:14:18.893 "state": "online", 00:14:18.893 "raid_level": "raid1", 00:14:18.893 "superblock": true, 00:14:18.893 "num_base_bdevs": 4, 00:14:18.893 "num_base_bdevs_discovered": 4, 00:14:18.893 "num_base_bdevs_operational": 4, 00:14:18.893 "base_bdevs_list": [ 00:14:18.893 { 00:14:18.893 "name": "BaseBdev1", 00:14:18.893 "uuid": "892181b7-65a5-5141-84be-a192a649bf64", 00:14:18.893 "is_configured": true, 00:14:18.893 "data_offset": 2048, 00:14:18.893 "data_size": 63488 00:14:18.893 }, 00:14:18.893 { 00:14:18.893 "name": "BaseBdev2", 00:14:18.893 "uuid": "69ca8ec4-53fc-5528-b91d-668febe69198", 00:14:18.893 "is_configured": true, 00:14:18.893 "data_offset": 2048, 00:14:18.893 "data_size": 63488 00:14:18.893 }, 00:14:18.893 { 00:14:18.893 "name": "BaseBdev3", 00:14:18.893 "uuid": "0c9a225c-51cb-52ba-8f5c-eba176eb7ad0", 00:14:18.893 "is_configured": true, 00:14:18.893 "data_offset": 2048, 00:14:18.893 "data_size": 63488 00:14:18.893 }, 00:14:18.893 { 00:14:18.893 "name": "BaseBdev4", 00:14:18.893 "uuid": "8fe4c29c-700e-52c1-a0ff-21d9f1e42fe8", 00:14:18.893 "is_configured": true, 00:14:18.893 "data_offset": 2048, 00:14:18.893 "data_size": 63488 00:14:18.893 } 00:14:18.893 ] 00:14:18.893 }' 00:14:18.893 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:18.893 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:19.151 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:14:19.151 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:19.151 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:19.151 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:19.151 [2024-11-19 15:20:09.465170] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:19.151 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:19.411 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:14:19.411 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:19.411 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:14:19.411 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:19.411 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:19.411 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:19.411 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:14:19.411 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:14:19.411 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:14:19.411 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:14:19.411 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:19.411 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:19.411 [2024-11-19 15:20:09.548691] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:19.411 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:19.411 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:14:19.411 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:19.411 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:19.411 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:19.411 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:19.411 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:19.411 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:19.411 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:19.411 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:19.411 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:19.411 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:19.411 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:19.411 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:19.411 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:19.411 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:19.411 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:19.411 "name": "raid_bdev1", 00:14:19.411 "uuid": "a426f079-d827-4102-96e5-ce4cc6c6d4c0", 00:14:19.411 "strip_size_kb": 0, 00:14:19.411 "state": "online", 00:14:19.411 "raid_level": "raid1", 00:14:19.411 "superblock": true, 00:14:19.411 "num_base_bdevs": 4, 00:14:19.411 "num_base_bdevs_discovered": 3, 00:14:19.411 "num_base_bdevs_operational": 3, 00:14:19.411 "base_bdevs_list": [ 00:14:19.411 { 00:14:19.411 "name": null, 00:14:19.411 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:19.411 "is_configured": false, 00:14:19.411 "data_offset": 0, 00:14:19.411 "data_size": 63488 00:14:19.411 }, 00:14:19.411 { 00:14:19.411 "name": "BaseBdev2", 00:14:19.411 "uuid": "69ca8ec4-53fc-5528-b91d-668febe69198", 00:14:19.411 "is_configured": true, 00:14:19.411 "data_offset": 2048, 00:14:19.411 "data_size": 63488 00:14:19.411 }, 00:14:19.411 { 00:14:19.411 "name": "BaseBdev3", 00:14:19.411 "uuid": "0c9a225c-51cb-52ba-8f5c-eba176eb7ad0", 00:14:19.411 "is_configured": true, 00:14:19.411 "data_offset": 2048, 00:14:19.411 "data_size": 63488 00:14:19.411 }, 00:14:19.411 { 00:14:19.411 "name": "BaseBdev4", 00:14:19.411 "uuid": "8fe4c29c-700e-52c1-a0ff-21d9f1e42fe8", 00:14:19.411 "is_configured": true, 00:14:19.411 "data_offset": 2048, 00:14:19.411 "data_size": 63488 00:14:19.411 } 00:14:19.411 ] 00:14:19.411 }' 00:14:19.411 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:19.411 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:19.411 [2024-11-19 15:20:09.635909] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:14:19.411 I/O size of 3145728 is greater than zero copy threshold (65536). 00:14:19.411 Zero copy mechanism will not be used. 00:14:19.411 Running I/O for 60 seconds... 00:14:19.671 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:19.671 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:19.671 15:20:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:19.671 [2024-11-19 15:20:09.998436] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:19.931 15:20:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:19.931 15:20:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:14:19.931 [2024-11-19 15:20:10.072849] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002940 00:14:19.931 [2024-11-19 15:20:10.074905] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:19.931 [2024-11-19 15:20:10.198124] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:14:19.931 [2024-11-19 15:20:10.198489] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:14:20.191 [2024-11-19 15:20:10.409514] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:14:20.191 [2024-11-19 15:20:10.410120] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:14:20.451 171.00 IOPS, 513.00 MiB/s [2024-11-19T15:20:10.790Z] [2024-11-19 15:20:10.775899] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:14:20.721 [2024-11-19 15:20:10.989391] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:14:20.721 [2024-11-19 15:20:10.989962] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:14:20.721 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:20.721 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:20.721 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:20.721 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:20.721 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:20.721 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:20.721 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:20.721 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:20.721 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:20.992 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:20.992 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:20.992 "name": "raid_bdev1", 00:14:20.992 "uuid": "a426f079-d827-4102-96e5-ce4cc6c6d4c0", 00:14:20.992 "strip_size_kb": 0, 00:14:20.992 "state": "online", 00:14:20.992 "raid_level": "raid1", 00:14:20.992 "superblock": true, 00:14:20.992 "num_base_bdevs": 4, 00:14:20.992 "num_base_bdevs_discovered": 4, 00:14:20.992 "num_base_bdevs_operational": 4, 00:14:20.992 "process": { 00:14:20.992 "type": "rebuild", 00:14:20.992 "target": "spare", 00:14:20.992 "progress": { 00:14:20.992 "blocks": 10240, 00:14:20.992 "percent": 16 00:14:20.992 } 00:14:20.992 }, 00:14:20.992 "base_bdevs_list": [ 00:14:20.992 { 00:14:20.992 "name": "spare", 00:14:20.992 "uuid": "1bfec8bc-e5c1-5e7b-b8b1-32c7631a4506", 00:14:20.992 "is_configured": true, 00:14:20.992 "data_offset": 2048, 00:14:20.992 "data_size": 63488 00:14:20.992 }, 00:14:20.992 { 00:14:20.992 "name": "BaseBdev2", 00:14:20.992 "uuid": "69ca8ec4-53fc-5528-b91d-668febe69198", 00:14:20.992 "is_configured": true, 00:14:20.992 "data_offset": 2048, 00:14:20.992 "data_size": 63488 00:14:20.992 }, 00:14:20.992 { 00:14:20.992 "name": "BaseBdev3", 00:14:20.992 "uuid": "0c9a225c-51cb-52ba-8f5c-eba176eb7ad0", 00:14:20.992 "is_configured": true, 00:14:20.992 "data_offset": 2048, 00:14:20.992 "data_size": 63488 00:14:20.992 }, 00:14:20.992 { 00:14:20.992 "name": "BaseBdev4", 00:14:20.992 "uuid": "8fe4c29c-700e-52c1-a0ff-21d9f1e42fe8", 00:14:20.992 "is_configured": true, 00:14:20.992 "data_offset": 2048, 00:14:20.992 "data_size": 63488 00:14:20.992 } 00:14:20.992 ] 00:14:20.992 }' 00:14:20.992 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:20.992 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:20.992 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:20.992 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:20.992 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:14:20.992 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:20.992 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:20.992 [2024-11-19 15:20:11.199337] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:20.992 [2024-11-19 15:20:11.268277] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:14:20.992 [2024-11-19 15:20:11.277746] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:20.992 [2024-11-19 15:20:11.277786] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:20.992 [2024-11-19 15:20:11.277800] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:14:20.992 [2024-11-19 15:20:11.289112] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000002870 00:14:20.992 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:20.992 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:14:20.992 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:20.992 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:20.992 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:20.992 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:20.992 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:20.992 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:20.992 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:20.992 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:20.992 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:20.992 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:20.992 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:20.992 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:20.992 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:21.252 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:21.252 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:21.252 "name": "raid_bdev1", 00:14:21.252 "uuid": "a426f079-d827-4102-96e5-ce4cc6c6d4c0", 00:14:21.252 "strip_size_kb": 0, 00:14:21.252 "state": "online", 00:14:21.252 "raid_level": "raid1", 00:14:21.252 "superblock": true, 00:14:21.252 "num_base_bdevs": 4, 00:14:21.252 "num_base_bdevs_discovered": 3, 00:14:21.252 "num_base_bdevs_operational": 3, 00:14:21.252 "base_bdevs_list": [ 00:14:21.252 { 00:14:21.252 "name": null, 00:14:21.252 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:21.252 "is_configured": false, 00:14:21.252 "data_offset": 0, 00:14:21.252 "data_size": 63488 00:14:21.252 }, 00:14:21.252 { 00:14:21.252 "name": "BaseBdev2", 00:14:21.252 "uuid": "69ca8ec4-53fc-5528-b91d-668febe69198", 00:14:21.252 "is_configured": true, 00:14:21.252 "data_offset": 2048, 00:14:21.252 "data_size": 63488 00:14:21.252 }, 00:14:21.252 { 00:14:21.252 "name": "BaseBdev3", 00:14:21.252 "uuid": "0c9a225c-51cb-52ba-8f5c-eba176eb7ad0", 00:14:21.252 "is_configured": true, 00:14:21.252 "data_offset": 2048, 00:14:21.252 "data_size": 63488 00:14:21.252 }, 00:14:21.252 { 00:14:21.252 "name": "BaseBdev4", 00:14:21.252 "uuid": "8fe4c29c-700e-52c1-a0ff-21d9f1e42fe8", 00:14:21.252 "is_configured": true, 00:14:21.252 "data_offset": 2048, 00:14:21.252 "data_size": 63488 00:14:21.252 } 00:14:21.252 ] 00:14:21.252 }' 00:14:21.252 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:21.252 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:21.512 162.00 IOPS, 486.00 MiB/s [2024-11-19T15:20:11.851Z] 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:21.512 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:21.512 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:21.512 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:21.512 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:21.512 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:21.512 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:21.512 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:21.512 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:21.512 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:21.512 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:21.512 "name": "raid_bdev1", 00:14:21.512 "uuid": "a426f079-d827-4102-96e5-ce4cc6c6d4c0", 00:14:21.512 "strip_size_kb": 0, 00:14:21.512 "state": "online", 00:14:21.512 "raid_level": "raid1", 00:14:21.512 "superblock": true, 00:14:21.512 "num_base_bdevs": 4, 00:14:21.512 "num_base_bdevs_discovered": 3, 00:14:21.512 "num_base_bdevs_operational": 3, 00:14:21.512 "base_bdevs_list": [ 00:14:21.512 { 00:14:21.512 "name": null, 00:14:21.512 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:21.512 "is_configured": false, 00:14:21.512 "data_offset": 0, 00:14:21.512 "data_size": 63488 00:14:21.512 }, 00:14:21.512 { 00:14:21.512 "name": "BaseBdev2", 00:14:21.512 "uuid": "69ca8ec4-53fc-5528-b91d-668febe69198", 00:14:21.512 "is_configured": true, 00:14:21.512 "data_offset": 2048, 00:14:21.512 "data_size": 63488 00:14:21.512 }, 00:14:21.512 { 00:14:21.512 "name": "BaseBdev3", 00:14:21.512 "uuid": "0c9a225c-51cb-52ba-8f5c-eba176eb7ad0", 00:14:21.512 "is_configured": true, 00:14:21.512 "data_offset": 2048, 00:14:21.512 "data_size": 63488 00:14:21.512 }, 00:14:21.512 { 00:14:21.512 "name": "BaseBdev4", 00:14:21.512 "uuid": "8fe4c29c-700e-52c1-a0ff-21d9f1e42fe8", 00:14:21.512 "is_configured": true, 00:14:21.512 "data_offset": 2048, 00:14:21.512 "data_size": 63488 00:14:21.512 } 00:14:21.512 ] 00:14:21.512 }' 00:14:21.512 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:21.512 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:21.512 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:21.773 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:21.773 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:21.773 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:21.773 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:21.773 [2024-11-19 15:20:11.877608] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:21.773 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:21.773 15:20:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:14:21.773 [2024-11-19 15:20:11.931816] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:14:21.773 [2024-11-19 15:20:11.933830] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:21.773 [2024-11-19 15:20:12.053740] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:14:21.773 [2024-11-19 15:20:12.054111] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:14:22.033 [2024-11-19 15:20:12.256270] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:14:22.033 [2024-11-19 15:20:12.256526] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:14:22.293 [2024-11-19 15:20:12.511567] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:14:22.293 [2024-11-19 15:20:12.517974] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:14:22.552 168.33 IOPS, 505.00 MiB/s [2024-11-19T15:20:12.891Z] [2024-11-19 15:20:12.746584] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:14:22.812 15:20:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:22.812 15:20:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:22.812 15:20:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:22.812 15:20:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:22.812 15:20:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:22.812 15:20:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:22.812 15:20:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:22.812 15:20:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:22.812 15:20:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:22.812 15:20:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:22.812 15:20:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:22.812 "name": "raid_bdev1", 00:14:22.812 "uuid": "a426f079-d827-4102-96e5-ce4cc6c6d4c0", 00:14:22.812 "strip_size_kb": 0, 00:14:22.812 "state": "online", 00:14:22.812 "raid_level": "raid1", 00:14:22.812 "superblock": true, 00:14:22.812 "num_base_bdevs": 4, 00:14:22.813 "num_base_bdevs_discovered": 4, 00:14:22.813 "num_base_bdevs_operational": 4, 00:14:22.813 "process": { 00:14:22.813 "type": "rebuild", 00:14:22.813 "target": "spare", 00:14:22.813 "progress": { 00:14:22.813 "blocks": 12288, 00:14:22.813 "percent": 19 00:14:22.813 } 00:14:22.813 }, 00:14:22.813 "base_bdevs_list": [ 00:14:22.813 { 00:14:22.813 "name": "spare", 00:14:22.813 "uuid": "1bfec8bc-e5c1-5e7b-b8b1-32c7631a4506", 00:14:22.813 "is_configured": true, 00:14:22.813 "data_offset": 2048, 00:14:22.813 "data_size": 63488 00:14:22.813 }, 00:14:22.813 { 00:14:22.813 "name": "BaseBdev2", 00:14:22.813 "uuid": "69ca8ec4-53fc-5528-b91d-668febe69198", 00:14:22.813 "is_configured": true, 00:14:22.813 "data_offset": 2048, 00:14:22.813 "data_size": 63488 00:14:22.813 }, 00:14:22.813 { 00:14:22.813 "name": "BaseBdev3", 00:14:22.813 "uuid": "0c9a225c-51cb-52ba-8f5c-eba176eb7ad0", 00:14:22.813 "is_configured": true, 00:14:22.813 "data_offset": 2048, 00:14:22.813 "data_size": 63488 00:14:22.813 }, 00:14:22.813 { 00:14:22.813 "name": "BaseBdev4", 00:14:22.813 "uuid": "8fe4c29c-700e-52c1-a0ff-21d9f1e42fe8", 00:14:22.813 "is_configured": true, 00:14:22.813 "data_offset": 2048, 00:14:22.813 "data_size": 63488 00:14:22.813 } 00:14:22.813 ] 00:14:22.813 }' 00:14:22.813 15:20:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:22.813 [2024-11-19 15:20:12.982407] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:14:22.813 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:22.813 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:22.813 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:22.813 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:14:22.813 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:14:22.813 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:14:22.813 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:14:22.813 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:14:22.813 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:14:22.813 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:14:22.813 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:22.813 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:22.813 [2024-11-19 15:20:13.076762] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:14:22.813 [2024-11-19 15:20:13.097165] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:14:22.813 [2024-11-19 15:20:13.097405] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:14:23.073 [2024-11-19 15:20:13.203480] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000002870 00:14:23.073 [2024-11-19 15:20:13.203544] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000002a10 00:14:23.073 [2024-11-19 15:20:13.204091] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:14:23.073 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:23.073 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:14:23.073 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:14:23.073 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:23.073 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:23.073 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:23.073 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:23.073 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:23.073 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:23.073 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:23.073 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:23.073 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:23.073 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:23.073 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:23.073 "name": "raid_bdev1", 00:14:23.073 "uuid": "a426f079-d827-4102-96e5-ce4cc6c6d4c0", 00:14:23.073 "strip_size_kb": 0, 00:14:23.073 "state": "online", 00:14:23.073 "raid_level": "raid1", 00:14:23.073 "superblock": true, 00:14:23.073 "num_base_bdevs": 4, 00:14:23.073 "num_base_bdevs_discovered": 3, 00:14:23.073 "num_base_bdevs_operational": 3, 00:14:23.073 "process": { 00:14:23.073 "type": "rebuild", 00:14:23.073 "target": "spare", 00:14:23.073 "progress": { 00:14:23.073 "blocks": 16384, 00:14:23.073 "percent": 25 00:14:23.073 } 00:14:23.073 }, 00:14:23.073 "base_bdevs_list": [ 00:14:23.073 { 00:14:23.073 "name": "spare", 00:14:23.073 "uuid": "1bfec8bc-e5c1-5e7b-b8b1-32c7631a4506", 00:14:23.073 "is_configured": true, 00:14:23.073 "data_offset": 2048, 00:14:23.073 "data_size": 63488 00:14:23.073 }, 00:14:23.073 { 00:14:23.073 "name": null, 00:14:23.073 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:23.073 "is_configured": false, 00:14:23.073 "data_offset": 0, 00:14:23.073 "data_size": 63488 00:14:23.073 }, 00:14:23.073 { 00:14:23.073 "name": "BaseBdev3", 00:14:23.073 "uuid": "0c9a225c-51cb-52ba-8f5c-eba176eb7ad0", 00:14:23.073 "is_configured": true, 00:14:23.073 "data_offset": 2048, 00:14:23.073 "data_size": 63488 00:14:23.073 }, 00:14:23.073 { 00:14:23.073 "name": "BaseBdev4", 00:14:23.073 "uuid": "8fe4c29c-700e-52c1-a0ff-21d9f1e42fe8", 00:14:23.073 "is_configured": true, 00:14:23.073 "data_offset": 2048, 00:14:23.073 "data_size": 63488 00:14:23.073 } 00:14:23.073 ] 00:14:23.073 }' 00:14:23.073 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:23.073 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:23.074 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:23.074 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:23.074 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@706 -- # local timeout=411 00:14:23.074 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:23.074 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:23.074 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:23.074 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:23.074 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:23.074 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:23.074 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:23.074 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:23.074 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:23.074 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:23.074 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:23.334 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:23.334 "name": "raid_bdev1", 00:14:23.334 "uuid": "a426f079-d827-4102-96e5-ce4cc6c6d4c0", 00:14:23.334 "strip_size_kb": 0, 00:14:23.334 "state": "online", 00:14:23.334 "raid_level": "raid1", 00:14:23.334 "superblock": true, 00:14:23.334 "num_base_bdevs": 4, 00:14:23.334 "num_base_bdevs_discovered": 3, 00:14:23.334 "num_base_bdevs_operational": 3, 00:14:23.334 "process": { 00:14:23.334 "type": "rebuild", 00:14:23.334 "target": "spare", 00:14:23.334 "progress": { 00:14:23.334 "blocks": 16384, 00:14:23.334 "percent": 25 00:14:23.334 } 00:14:23.334 }, 00:14:23.334 "base_bdevs_list": [ 00:14:23.334 { 00:14:23.334 "name": "spare", 00:14:23.334 "uuid": "1bfec8bc-e5c1-5e7b-b8b1-32c7631a4506", 00:14:23.334 "is_configured": true, 00:14:23.334 "data_offset": 2048, 00:14:23.334 "data_size": 63488 00:14:23.334 }, 00:14:23.334 { 00:14:23.334 "name": null, 00:14:23.334 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:23.334 "is_configured": false, 00:14:23.334 "data_offset": 0, 00:14:23.334 "data_size": 63488 00:14:23.334 }, 00:14:23.334 { 00:14:23.334 "name": "BaseBdev3", 00:14:23.334 "uuid": "0c9a225c-51cb-52ba-8f5c-eba176eb7ad0", 00:14:23.334 "is_configured": true, 00:14:23.334 "data_offset": 2048, 00:14:23.334 "data_size": 63488 00:14:23.334 }, 00:14:23.334 { 00:14:23.334 "name": "BaseBdev4", 00:14:23.334 "uuid": "8fe4c29c-700e-52c1-a0ff-21d9f1e42fe8", 00:14:23.334 "is_configured": true, 00:14:23.334 "data_offset": 2048, 00:14:23.334 "data_size": 63488 00:14:23.334 } 00:14:23.334 ] 00:14:23.334 }' 00:14:23.334 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:23.334 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:23.334 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:23.334 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:23.334 15:20:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:23.334 [2024-11-19 15:20:13.508325] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:14:23.334 [2024-11-19 15:20:13.508814] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:14:23.904 141.25 IOPS, 423.75 MiB/s [2024-11-19T15:20:14.243Z] [2024-11-19 15:20:13.983059] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 26624 offset_begin: 24576 offset_end: 30720 00:14:23.904 [2024-11-19 15:20:14.214491] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 28672 offset_begin: 24576 offset_end: 30720 00:14:24.164 [2024-11-19 15:20:14.449232] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 32768 offset_begin: 30720 offset_end: 36864 00:14:24.164 15:20:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:24.164 15:20:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:24.164 15:20:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:24.164 15:20:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:24.164 15:20:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:24.164 15:20:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:24.425 15:20:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:24.425 15:20:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:24.425 15:20:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:24.425 15:20:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:24.425 15:20:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:24.425 15:20:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:24.425 "name": "raid_bdev1", 00:14:24.425 "uuid": "a426f079-d827-4102-96e5-ce4cc6c6d4c0", 00:14:24.425 "strip_size_kb": 0, 00:14:24.425 "state": "online", 00:14:24.425 "raid_level": "raid1", 00:14:24.425 "superblock": true, 00:14:24.425 "num_base_bdevs": 4, 00:14:24.425 "num_base_bdevs_discovered": 3, 00:14:24.425 "num_base_bdevs_operational": 3, 00:14:24.425 "process": { 00:14:24.425 "type": "rebuild", 00:14:24.425 "target": "spare", 00:14:24.425 "progress": { 00:14:24.425 "blocks": 32768, 00:14:24.425 "percent": 51 00:14:24.425 } 00:14:24.425 }, 00:14:24.425 "base_bdevs_list": [ 00:14:24.425 { 00:14:24.425 "name": "spare", 00:14:24.425 "uuid": "1bfec8bc-e5c1-5e7b-b8b1-32c7631a4506", 00:14:24.425 "is_configured": true, 00:14:24.425 "data_offset": 2048, 00:14:24.425 "data_size": 63488 00:14:24.425 }, 00:14:24.425 { 00:14:24.425 "name": null, 00:14:24.425 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:24.425 "is_configured": false, 00:14:24.425 "data_offset": 0, 00:14:24.425 "data_size": 63488 00:14:24.425 }, 00:14:24.425 { 00:14:24.425 "name": "BaseBdev3", 00:14:24.425 "uuid": "0c9a225c-51cb-52ba-8f5c-eba176eb7ad0", 00:14:24.425 "is_configured": true, 00:14:24.425 "data_offset": 2048, 00:14:24.425 "data_size": 63488 00:14:24.425 }, 00:14:24.425 { 00:14:24.425 "name": "BaseBdev4", 00:14:24.425 "uuid": "8fe4c29c-700e-52c1-a0ff-21d9f1e42fe8", 00:14:24.425 "is_configured": true, 00:14:24.425 "data_offset": 2048, 00:14:24.425 "data_size": 63488 00:14:24.425 } 00:14:24.425 ] 00:14:24.425 }' 00:14:24.425 15:20:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:24.425 [2024-11-19 15:20:14.582483] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 34816 offset_begin: 30720 offset_end: 36864 00:14:24.425 15:20:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:24.425 15:20:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:24.425 122.40 IOPS, 367.20 MiB/s [2024-11-19T15:20:14.764Z] 15:20:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:24.425 15:20:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:24.685 [2024-11-19 15:20:14.916201] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 38912 offset_begin: 36864 offset_end: 43008 00:14:24.945 [2024-11-19 15:20:15.042454] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 40960 offset_begin: 36864 offset_end: 43008 00:14:25.205 [2024-11-19 15:20:15.470898] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 47104 offset_begin: 43008 offset_end: 49152 00:14:25.465 109.67 IOPS, 329.00 MiB/s [2024-11-19T15:20:15.804Z] 15:20:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:25.465 15:20:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:25.465 15:20:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:25.465 15:20:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:25.465 15:20:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:25.465 15:20:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:25.465 15:20:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:25.465 15:20:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:25.465 15:20:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:25.465 15:20:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:25.465 15:20:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:25.465 15:20:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:25.465 "name": "raid_bdev1", 00:14:25.465 "uuid": "a426f079-d827-4102-96e5-ce4cc6c6d4c0", 00:14:25.465 "strip_size_kb": 0, 00:14:25.465 "state": "online", 00:14:25.465 "raid_level": "raid1", 00:14:25.465 "superblock": true, 00:14:25.465 "num_base_bdevs": 4, 00:14:25.465 "num_base_bdevs_discovered": 3, 00:14:25.465 "num_base_bdevs_operational": 3, 00:14:25.465 "process": { 00:14:25.465 "type": "rebuild", 00:14:25.465 "target": "spare", 00:14:25.465 "progress": { 00:14:25.465 "blocks": 49152, 00:14:25.465 "percent": 77 00:14:25.465 } 00:14:25.465 }, 00:14:25.465 "base_bdevs_list": [ 00:14:25.465 { 00:14:25.465 "name": "spare", 00:14:25.465 "uuid": "1bfec8bc-e5c1-5e7b-b8b1-32c7631a4506", 00:14:25.465 "is_configured": true, 00:14:25.465 "data_offset": 2048, 00:14:25.465 "data_size": 63488 00:14:25.465 }, 00:14:25.465 { 00:14:25.465 "name": null, 00:14:25.465 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:25.465 "is_configured": false, 00:14:25.465 "data_offset": 0, 00:14:25.466 "data_size": 63488 00:14:25.466 }, 00:14:25.466 { 00:14:25.466 "name": "BaseBdev3", 00:14:25.466 "uuid": "0c9a225c-51cb-52ba-8f5c-eba176eb7ad0", 00:14:25.466 "is_configured": true, 00:14:25.466 "data_offset": 2048, 00:14:25.466 "data_size": 63488 00:14:25.466 }, 00:14:25.466 { 00:14:25.466 "name": "BaseBdev4", 00:14:25.466 "uuid": "8fe4c29c-700e-52c1-a0ff-21d9f1e42fe8", 00:14:25.466 "is_configured": true, 00:14:25.466 "data_offset": 2048, 00:14:25.466 "data_size": 63488 00:14:25.466 } 00:14:25.466 ] 00:14:25.466 }' 00:14:25.466 15:20:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:25.466 15:20:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:25.466 15:20:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:25.466 [2024-11-19 15:20:15.793452] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 53248 offset_begin: 49152 offset_end: 55296 00:14:25.725 15:20:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:25.725 15:20:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:25.725 [2024-11-19 15:20:16.008093] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 57344 offset_begin: 55296 offset_end: 61440 00:14:26.295 [2024-11-19 15:20:16.335400] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:14:26.296 [2024-11-19 15:20:16.435250] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:14:26.296 [2024-11-19 15:20:16.437502] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:26.556 99.14 IOPS, 297.43 MiB/s [2024-11-19T15:20:16.895Z] 15:20:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:26.556 15:20:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:26.556 15:20:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:26.556 15:20:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:26.556 15:20:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:26.556 15:20:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:26.556 15:20:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:26.556 15:20:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:26.556 15:20:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.556 15:20:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:26.556 15:20:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.556 15:20:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:26.556 "name": "raid_bdev1", 00:14:26.556 "uuid": "a426f079-d827-4102-96e5-ce4cc6c6d4c0", 00:14:26.556 "strip_size_kb": 0, 00:14:26.556 "state": "online", 00:14:26.556 "raid_level": "raid1", 00:14:26.556 "superblock": true, 00:14:26.556 "num_base_bdevs": 4, 00:14:26.556 "num_base_bdevs_discovered": 3, 00:14:26.556 "num_base_bdevs_operational": 3, 00:14:26.556 "base_bdevs_list": [ 00:14:26.556 { 00:14:26.556 "name": "spare", 00:14:26.556 "uuid": "1bfec8bc-e5c1-5e7b-b8b1-32c7631a4506", 00:14:26.556 "is_configured": true, 00:14:26.556 "data_offset": 2048, 00:14:26.556 "data_size": 63488 00:14:26.556 }, 00:14:26.556 { 00:14:26.556 "name": null, 00:14:26.556 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:26.556 "is_configured": false, 00:14:26.556 "data_offset": 0, 00:14:26.556 "data_size": 63488 00:14:26.556 }, 00:14:26.556 { 00:14:26.556 "name": "BaseBdev3", 00:14:26.556 "uuid": "0c9a225c-51cb-52ba-8f5c-eba176eb7ad0", 00:14:26.556 "is_configured": true, 00:14:26.556 "data_offset": 2048, 00:14:26.556 "data_size": 63488 00:14:26.556 }, 00:14:26.556 { 00:14:26.556 "name": "BaseBdev4", 00:14:26.556 "uuid": "8fe4c29c-700e-52c1-a0ff-21d9f1e42fe8", 00:14:26.556 "is_configured": true, 00:14:26.556 "data_offset": 2048, 00:14:26.556 "data_size": 63488 00:14:26.556 } 00:14:26.556 ] 00:14:26.556 }' 00:14:26.556 15:20:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:26.817 15:20:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:14:26.817 15:20:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:26.817 15:20:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:14:26.817 15:20:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@709 -- # break 00:14:26.817 15:20:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:26.817 15:20:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:26.817 15:20:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:26.817 15:20:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:26.817 15:20:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:26.817 15:20:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:26.817 15:20:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:26.817 15:20:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.817 15:20:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:26.817 15:20:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.817 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:26.817 "name": "raid_bdev1", 00:14:26.817 "uuid": "a426f079-d827-4102-96e5-ce4cc6c6d4c0", 00:14:26.817 "strip_size_kb": 0, 00:14:26.817 "state": "online", 00:14:26.817 "raid_level": "raid1", 00:14:26.817 "superblock": true, 00:14:26.817 "num_base_bdevs": 4, 00:14:26.817 "num_base_bdevs_discovered": 3, 00:14:26.817 "num_base_bdevs_operational": 3, 00:14:26.817 "base_bdevs_list": [ 00:14:26.817 { 00:14:26.817 "name": "spare", 00:14:26.817 "uuid": "1bfec8bc-e5c1-5e7b-b8b1-32c7631a4506", 00:14:26.817 "is_configured": true, 00:14:26.817 "data_offset": 2048, 00:14:26.817 "data_size": 63488 00:14:26.817 }, 00:14:26.817 { 00:14:26.817 "name": null, 00:14:26.817 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:26.817 "is_configured": false, 00:14:26.817 "data_offset": 0, 00:14:26.817 "data_size": 63488 00:14:26.817 }, 00:14:26.817 { 00:14:26.817 "name": "BaseBdev3", 00:14:26.817 "uuid": "0c9a225c-51cb-52ba-8f5c-eba176eb7ad0", 00:14:26.817 "is_configured": true, 00:14:26.817 "data_offset": 2048, 00:14:26.817 "data_size": 63488 00:14:26.817 }, 00:14:26.817 { 00:14:26.817 "name": "BaseBdev4", 00:14:26.817 "uuid": "8fe4c29c-700e-52c1-a0ff-21d9f1e42fe8", 00:14:26.817 "is_configured": true, 00:14:26.817 "data_offset": 2048, 00:14:26.817 "data_size": 63488 00:14:26.817 } 00:14:26.817 ] 00:14:26.817 }' 00:14:26.817 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:26.817 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:26.817 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:26.817 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:26.817 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:14:26.817 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:26.817 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:26.817 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:26.817 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:26.817 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:26.817 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:26.817 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:26.817 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:26.817 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:26.817 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:26.817 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:26.817 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.817 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:26.817 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.817 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:26.817 "name": "raid_bdev1", 00:14:26.817 "uuid": "a426f079-d827-4102-96e5-ce4cc6c6d4c0", 00:14:26.817 "strip_size_kb": 0, 00:14:26.817 "state": "online", 00:14:26.817 "raid_level": "raid1", 00:14:26.817 "superblock": true, 00:14:26.817 "num_base_bdevs": 4, 00:14:26.817 "num_base_bdevs_discovered": 3, 00:14:26.817 "num_base_bdevs_operational": 3, 00:14:26.817 "base_bdevs_list": [ 00:14:26.817 { 00:14:26.817 "name": "spare", 00:14:26.817 "uuid": "1bfec8bc-e5c1-5e7b-b8b1-32c7631a4506", 00:14:26.817 "is_configured": true, 00:14:26.817 "data_offset": 2048, 00:14:26.817 "data_size": 63488 00:14:26.817 }, 00:14:26.817 { 00:14:26.817 "name": null, 00:14:26.817 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:26.817 "is_configured": false, 00:14:26.817 "data_offset": 0, 00:14:26.817 "data_size": 63488 00:14:26.817 }, 00:14:26.817 { 00:14:26.817 "name": "BaseBdev3", 00:14:26.817 "uuid": "0c9a225c-51cb-52ba-8f5c-eba176eb7ad0", 00:14:26.817 "is_configured": true, 00:14:26.817 "data_offset": 2048, 00:14:26.817 "data_size": 63488 00:14:26.817 }, 00:14:26.817 { 00:14:26.817 "name": "BaseBdev4", 00:14:26.817 "uuid": "8fe4c29c-700e-52c1-a0ff-21d9f1e42fe8", 00:14:26.817 "is_configured": true, 00:14:26.817 "data_offset": 2048, 00:14:26.817 "data_size": 63488 00:14:26.817 } 00:14:26.817 ] 00:14:26.817 }' 00:14:26.817 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:26.817 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:27.387 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:27.387 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:27.387 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:27.387 [2024-11-19 15:20:17.555115] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:27.387 [2024-11-19 15:20:17.555150] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:27.387 00:14:27.387 Latency(us) 00:14:27.387 [2024-11-19T15:20:17.726Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:14:27.387 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:14:27.387 raid_bdev1 : 8.00 90.79 272.38 0.00 0.00 15548.92 289.76 115389.15 00:14:27.387 [2024-11-19T15:20:17.726Z] =================================================================================================================== 00:14:27.387 [2024-11-19T15:20:17.726Z] Total : 90.79 272.38 0.00 0.00 15548.92 289.76 115389.15 00:14:27.387 { 00:14:27.387 "results": [ 00:14:27.387 { 00:14:27.387 "job": "raid_bdev1", 00:14:27.387 "core_mask": "0x1", 00:14:27.387 "workload": "randrw", 00:14:27.387 "percentage": 50, 00:14:27.387 "status": "finished", 00:14:27.387 "queue_depth": 2, 00:14:27.387 "io_size": 3145728, 00:14:27.387 "runtime": 7.99611, 00:14:27.387 "iops": 90.79414865478338, 00:14:27.387 "mibps": 272.38244596435015, 00:14:27.387 "io_failed": 0, 00:14:27.387 "io_timeout": 0, 00:14:27.387 "avg_latency_us": 15548.916525316683, 00:14:27.387 "min_latency_us": 289.7606986899563, 00:14:27.387 "max_latency_us": 115389.14934497817 00:14:27.387 } 00:14:27.387 ], 00:14:27.387 "core_count": 1 00:14:27.387 } 00:14:27.387 [2024-11-19 15:20:17.622013] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:27.387 [2024-11-19 15:20:17.622060] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:27.387 [2024-11-19 15:20:17.622158] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:27.387 [2024-11-19 15:20:17.622170] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:14:27.387 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:27.387 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:27.387 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:27.387 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # jq length 00:14:27.387 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:27.387 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:27.387 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:14:27.387 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:14:27.387 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:14:27.387 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:14:27.387 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:14:27.387 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:14:27.387 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:14:27.387 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:14:27.387 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:14:27.387 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:14:27.387 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:14:27.387 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:27.387 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:14:27.647 /dev/nbd0 00:14:27.647 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:14:27.647 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:14:27.647 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:14:27.647 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # local i 00:14:27.647 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:14:27.647 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:14:27.647 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:14:27.647 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@877 -- # break 00:14:27.647 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:14:27.647 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:14:27.647 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:27.647 1+0 records in 00:14:27.647 1+0 records out 00:14:27.647 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000272006 s, 15.1 MB/s 00:14:27.648 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:27.648 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # size=4096 00:14:27.648 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:27.648 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:14:27.648 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@893 -- # return 0 00:14:27.648 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:27.648 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:27.648 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:14:27.648 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z '' ']' 00:14:27.648 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@728 -- # continue 00:14:27.648 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:14:27.648 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev3 ']' 00:14:27.648 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev3 /dev/nbd1 00:14:27.648 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:14:27.648 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev3') 00:14:27.648 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:14:27.648 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:14:27.648 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:14:27.648 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:14:27.648 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:14:27.648 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:27.648 15:20:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev3 /dev/nbd1 00:14:27.908 /dev/nbd1 00:14:27.908 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:14:27.908 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:14:27.908 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:14:27.908 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # local i 00:14:27.908 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:14:27.908 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:14:27.908 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:14:27.908 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@877 -- # break 00:14:27.908 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:14:27.908 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:14:27.908 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:27.908 1+0 records in 00:14:27.908 1+0 records out 00:14:27.908 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000481545 s, 8.5 MB/s 00:14:27.908 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:27.909 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # size=4096 00:14:27.909 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:27.909 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:14:27.909 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@893 -- # return 0 00:14:27.909 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:27.909 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:27.909 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@731 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:14:28.169 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:14:28.169 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:14:28.169 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:14:28.169 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:14:28.169 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:14:28.169 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:28.169 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:14:28.169 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:14:28.169 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:14:28.169 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:14:28.169 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:28.169 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:28.169 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:14:28.169 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:14:28.169 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:14:28.169 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:14:28.169 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev4 ']' 00:14:28.169 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev4 /dev/nbd1 00:14:28.169 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:14:28.169 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev4') 00:14:28.169 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:14:28.169 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:14:28.169 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:14:28.169 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:14:28.169 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:14:28.169 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:28.169 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev4 /dev/nbd1 00:14:28.429 /dev/nbd1 00:14:28.429 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:14:28.429 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:14:28.429 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:14:28.429 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # local i 00:14:28.429 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:14:28.429 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:14:28.429 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:14:28.429 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@877 -- # break 00:14:28.429 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:14:28.429 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:14:28.429 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:28.429 1+0 records in 00:14:28.429 1+0 records out 00:14:28.429 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000413934 s, 9.9 MB/s 00:14:28.429 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:28.429 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # size=4096 00:14:28.429 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:28.429 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:14:28.429 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@893 -- # return 0 00:14:28.429 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:28.429 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:28.429 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@731 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:14:28.429 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:14:28.429 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:14:28.429 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:14:28.429 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:14:28.429 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:14:28.429 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:28.429 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:14:28.690 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:14:28.690 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:14:28.690 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:14:28.690 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:28.690 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:28.690 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:14:28.690 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:14:28.690 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:14:28.690 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:14:28.690 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:14:28.690 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:14:28.690 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:14:28.690 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:14:28.690 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:28.690 15:20:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:14:28.951 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:14:28.951 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:14:28.951 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:14:28.951 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:28.951 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:28.951 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:14:28.951 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:14:28.951 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:14:28.951 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:14:28.951 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:14:28.951 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:28.951 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:28.951 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:28.951 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:14:28.951 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:28.951 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:28.951 [2024-11-19 15:20:19.166519] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:14:28.951 [2024-11-19 15:20:19.166638] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:28.951 [2024-11-19 15:20:19.166677] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:14:28.951 [2024-11-19 15:20:19.166710] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:28.951 [2024-11-19 15:20:19.168901] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:28.951 [2024-11-19 15:20:19.168981] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:14:28.951 [2024-11-19 15:20:19.169102] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:14:28.951 [2024-11-19 15:20:19.169161] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:28.951 [2024-11-19 15:20:19.169304] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:28.951 [2024-11-19 15:20:19.169432] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:14:28.951 spare 00:14:28.951 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:28.951 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:14:28.951 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:28.951 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:28.951 [2024-11-19 15:20:19.269348] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:14:28.951 [2024-11-19 15:20:19.269413] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:14:28.951 [2024-11-19 15:20:19.269684] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000337b0 00:14:28.951 [2024-11-19 15:20:19.269848] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:14:28.951 [2024-11-19 15:20:19.269891] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:14:28.951 [2024-11-19 15:20:19.270074] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:28.951 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:28.951 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:14:28.951 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:28.951 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:28.951 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:28.951 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:28.951 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:28.951 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:28.951 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:28.951 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:28.951 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:28.951 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:28.951 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:28.951 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:28.951 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:29.211 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:29.211 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:29.211 "name": "raid_bdev1", 00:14:29.211 "uuid": "a426f079-d827-4102-96e5-ce4cc6c6d4c0", 00:14:29.211 "strip_size_kb": 0, 00:14:29.211 "state": "online", 00:14:29.211 "raid_level": "raid1", 00:14:29.211 "superblock": true, 00:14:29.211 "num_base_bdevs": 4, 00:14:29.211 "num_base_bdevs_discovered": 3, 00:14:29.211 "num_base_bdevs_operational": 3, 00:14:29.211 "base_bdevs_list": [ 00:14:29.211 { 00:14:29.211 "name": "spare", 00:14:29.211 "uuid": "1bfec8bc-e5c1-5e7b-b8b1-32c7631a4506", 00:14:29.211 "is_configured": true, 00:14:29.211 "data_offset": 2048, 00:14:29.211 "data_size": 63488 00:14:29.211 }, 00:14:29.211 { 00:14:29.211 "name": null, 00:14:29.211 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:29.211 "is_configured": false, 00:14:29.211 "data_offset": 2048, 00:14:29.211 "data_size": 63488 00:14:29.212 }, 00:14:29.212 { 00:14:29.212 "name": "BaseBdev3", 00:14:29.212 "uuid": "0c9a225c-51cb-52ba-8f5c-eba176eb7ad0", 00:14:29.212 "is_configured": true, 00:14:29.212 "data_offset": 2048, 00:14:29.212 "data_size": 63488 00:14:29.212 }, 00:14:29.212 { 00:14:29.212 "name": "BaseBdev4", 00:14:29.212 "uuid": "8fe4c29c-700e-52c1-a0ff-21d9f1e42fe8", 00:14:29.212 "is_configured": true, 00:14:29.212 "data_offset": 2048, 00:14:29.212 "data_size": 63488 00:14:29.212 } 00:14:29.212 ] 00:14:29.212 }' 00:14:29.212 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:29.212 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:29.472 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:29.472 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:29.472 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:29.472 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:29.472 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:29.472 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:29.472 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:29.472 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:29.472 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:29.472 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:29.472 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:29.472 "name": "raid_bdev1", 00:14:29.472 "uuid": "a426f079-d827-4102-96e5-ce4cc6c6d4c0", 00:14:29.472 "strip_size_kb": 0, 00:14:29.472 "state": "online", 00:14:29.472 "raid_level": "raid1", 00:14:29.472 "superblock": true, 00:14:29.472 "num_base_bdevs": 4, 00:14:29.472 "num_base_bdevs_discovered": 3, 00:14:29.472 "num_base_bdevs_operational": 3, 00:14:29.472 "base_bdevs_list": [ 00:14:29.472 { 00:14:29.472 "name": "spare", 00:14:29.472 "uuid": "1bfec8bc-e5c1-5e7b-b8b1-32c7631a4506", 00:14:29.472 "is_configured": true, 00:14:29.472 "data_offset": 2048, 00:14:29.472 "data_size": 63488 00:14:29.472 }, 00:14:29.472 { 00:14:29.472 "name": null, 00:14:29.472 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:29.472 "is_configured": false, 00:14:29.472 "data_offset": 2048, 00:14:29.472 "data_size": 63488 00:14:29.472 }, 00:14:29.472 { 00:14:29.472 "name": "BaseBdev3", 00:14:29.472 "uuid": "0c9a225c-51cb-52ba-8f5c-eba176eb7ad0", 00:14:29.472 "is_configured": true, 00:14:29.472 "data_offset": 2048, 00:14:29.472 "data_size": 63488 00:14:29.472 }, 00:14:29.472 { 00:14:29.472 "name": "BaseBdev4", 00:14:29.472 "uuid": "8fe4c29c-700e-52c1-a0ff-21d9f1e42fe8", 00:14:29.472 "is_configured": true, 00:14:29.472 "data_offset": 2048, 00:14:29.472 "data_size": 63488 00:14:29.472 } 00:14:29.472 ] 00:14:29.472 }' 00:14:29.472 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:29.732 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:29.732 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:29.732 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:29.732 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:29.732 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:29.732 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:29.732 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:14:29.732 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:29.732 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:14:29.732 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:14:29.732 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:29.732 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:29.732 [2024-11-19 15:20:19.913425] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:29.732 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:29.732 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:14:29.732 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:29.732 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:29.732 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:29.732 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:29.733 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:29.733 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:29.733 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:29.733 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:29.733 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:29.733 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:29.733 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:29.733 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:29.733 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:29.733 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:29.733 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:29.733 "name": "raid_bdev1", 00:14:29.733 "uuid": "a426f079-d827-4102-96e5-ce4cc6c6d4c0", 00:14:29.733 "strip_size_kb": 0, 00:14:29.733 "state": "online", 00:14:29.733 "raid_level": "raid1", 00:14:29.733 "superblock": true, 00:14:29.733 "num_base_bdevs": 4, 00:14:29.733 "num_base_bdevs_discovered": 2, 00:14:29.733 "num_base_bdevs_operational": 2, 00:14:29.733 "base_bdevs_list": [ 00:14:29.733 { 00:14:29.733 "name": null, 00:14:29.733 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:29.733 "is_configured": false, 00:14:29.733 "data_offset": 0, 00:14:29.733 "data_size": 63488 00:14:29.733 }, 00:14:29.733 { 00:14:29.733 "name": null, 00:14:29.733 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:29.733 "is_configured": false, 00:14:29.733 "data_offset": 2048, 00:14:29.733 "data_size": 63488 00:14:29.733 }, 00:14:29.733 { 00:14:29.733 "name": "BaseBdev3", 00:14:29.733 "uuid": "0c9a225c-51cb-52ba-8f5c-eba176eb7ad0", 00:14:29.733 "is_configured": true, 00:14:29.733 "data_offset": 2048, 00:14:29.733 "data_size": 63488 00:14:29.733 }, 00:14:29.733 { 00:14:29.733 "name": "BaseBdev4", 00:14:29.733 "uuid": "8fe4c29c-700e-52c1-a0ff-21d9f1e42fe8", 00:14:29.733 "is_configured": true, 00:14:29.733 "data_offset": 2048, 00:14:29.733 "data_size": 63488 00:14:29.733 } 00:14:29.733 ] 00:14:29.733 }' 00:14:29.733 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:29.733 15:20:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:30.303 15:20:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:30.303 15:20:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:30.303 15:20:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:30.303 [2024-11-19 15:20:20.384694] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:30.303 [2024-11-19 15:20:20.384901] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:14:30.303 [2024-11-19 15:20:20.384973] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:14:30.303 [2024-11-19 15:20:20.385031] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:30.303 [2024-11-19 15:20:20.389587] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000033880 00:14:30.303 15:20:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:30.303 15:20:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@757 -- # sleep 1 00:14:30.303 [2024-11-19 15:20:20.391459] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:31.244 15:20:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:31.244 15:20:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:31.244 15:20:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:31.244 15:20:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:31.244 15:20:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:31.244 15:20:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:31.244 15:20:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:31.244 15:20:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:31.244 15:20:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:31.244 15:20:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:31.244 15:20:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:31.244 "name": "raid_bdev1", 00:14:31.244 "uuid": "a426f079-d827-4102-96e5-ce4cc6c6d4c0", 00:14:31.244 "strip_size_kb": 0, 00:14:31.244 "state": "online", 00:14:31.244 "raid_level": "raid1", 00:14:31.244 "superblock": true, 00:14:31.244 "num_base_bdevs": 4, 00:14:31.244 "num_base_bdevs_discovered": 3, 00:14:31.244 "num_base_bdevs_operational": 3, 00:14:31.244 "process": { 00:14:31.244 "type": "rebuild", 00:14:31.244 "target": "spare", 00:14:31.244 "progress": { 00:14:31.244 "blocks": 20480, 00:14:31.244 "percent": 32 00:14:31.244 } 00:14:31.244 }, 00:14:31.244 "base_bdevs_list": [ 00:14:31.244 { 00:14:31.244 "name": "spare", 00:14:31.244 "uuid": "1bfec8bc-e5c1-5e7b-b8b1-32c7631a4506", 00:14:31.244 "is_configured": true, 00:14:31.244 "data_offset": 2048, 00:14:31.244 "data_size": 63488 00:14:31.244 }, 00:14:31.244 { 00:14:31.244 "name": null, 00:14:31.244 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:31.244 "is_configured": false, 00:14:31.244 "data_offset": 2048, 00:14:31.244 "data_size": 63488 00:14:31.244 }, 00:14:31.244 { 00:14:31.244 "name": "BaseBdev3", 00:14:31.244 "uuid": "0c9a225c-51cb-52ba-8f5c-eba176eb7ad0", 00:14:31.244 "is_configured": true, 00:14:31.244 "data_offset": 2048, 00:14:31.244 "data_size": 63488 00:14:31.244 }, 00:14:31.244 { 00:14:31.244 "name": "BaseBdev4", 00:14:31.244 "uuid": "8fe4c29c-700e-52c1-a0ff-21d9f1e42fe8", 00:14:31.244 "is_configured": true, 00:14:31.244 "data_offset": 2048, 00:14:31.244 "data_size": 63488 00:14:31.244 } 00:14:31.244 ] 00:14:31.244 }' 00:14:31.244 15:20:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:31.244 15:20:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:31.244 15:20:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:31.244 15:20:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:31.244 15:20:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:14:31.244 15:20:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:31.244 15:20:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:31.244 [2024-11-19 15:20:21.555648] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:31.504 [2024-11-19 15:20:21.595400] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:14:31.504 [2024-11-19 15:20:21.595468] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:31.504 [2024-11-19 15:20:21.595484] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:31.504 [2024-11-19 15:20:21.595493] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:14:31.505 15:20:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:31.505 15:20:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:14:31.505 15:20:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:31.505 15:20:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:31.505 15:20:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:31.505 15:20:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:31.505 15:20:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:31.505 15:20:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:31.505 15:20:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:31.505 15:20:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:31.505 15:20:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:31.505 15:20:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:31.505 15:20:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:31.505 15:20:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:31.505 15:20:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:31.505 15:20:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:31.505 15:20:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:31.505 "name": "raid_bdev1", 00:14:31.505 "uuid": "a426f079-d827-4102-96e5-ce4cc6c6d4c0", 00:14:31.505 "strip_size_kb": 0, 00:14:31.505 "state": "online", 00:14:31.505 "raid_level": "raid1", 00:14:31.505 "superblock": true, 00:14:31.505 "num_base_bdevs": 4, 00:14:31.505 "num_base_bdevs_discovered": 2, 00:14:31.505 "num_base_bdevs_operational": 2, 00:14:31.505 "base_bdevs_list": [ 00:14:31.505 { 00:14:31.505 "name": null, 00:14:31.505 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:31.505 "is_configured": false, 00:14:31.505 "data_offset": 0, 00:14:31.505 "data_size": 63488 00:14:31.505 }, 00:14:31.505 { 00:14:31.505 "name": null, 00:14:31.505 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:31.505 "is_configured": false, 00:14:31.505 "data_offset": 2048, 00:14:31.505 "data_size": 63488 00:14:31.505 }, 00:14:31.505 { 00:14:31.505 "name": "BaseBdev3", 00:14:31.505 "uuid": "0c9a225c-51cb-52ba-8f5c-eba176eb7ad0", 00:14:31.505 "is_configured": true, 00:14:31.505 "data_offset": 2048, 00:14:31.505 "data_size": 63488 00:14:31.505 }, 00:14:31.505 { 00:14:31.505 "name": "BaseBdev4", 00:14:31.505 "uuid": "8fe4c29c-700e-52c1-a0ff-21d9f1e42fe8", 00:14:31.505 "is_configured": true, 00:14:31.505 "data_offset": 2048, 00:14:31.505 "data_size": 63488 00:14:31.505 } 00:14:31.505 ] 00:14:31.505 }' 00:14:31.505 15:20:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:31.505 15:20:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:32.075 15:20:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:14:32.075 15:20:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:32.075 15:20:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:32.075 [2024-11-19 15:20:22.110996] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:14:32.075 [2024-11-19 15:20:22.111105] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:32.075 [2024-11-19 15:20:22.111149] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ba80 00:14:32.075 [2024-11-19 15:20:22.111191] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:32.075 [2024-11-19 15:20:22.111609] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:32.075 [2024-11-19 15:20:22.111674] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:14:32.075 [2024-11-19 15:20:22.111790] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:14:32.075 [2024-11-19 15:20:22.111849] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:14:32.075 [2024-11-19 15:20:22.111889] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:14:32.075 [2024-11-19 15:20:22.111989] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:32.075 [2024-11-19 15:20:22.115819] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000033950 00:14:32.075 spare 00:14:32.075 15:20:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:32.075 15:20:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@764 -- # sleep 1 00:14:32.075 [2024-11-19 15:20:22.117749] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:33.017 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:33.017 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:33.017 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:33.017 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:33.017 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:33.017 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:33.017 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:33.017 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:33.017 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:33.017 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:33.017 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:33.017 "name": "raid_bdev1", 00:14:33.017 "uuid": "a426f079-d827-4102-96e5-ce4cc6c6d4c0", 00:14:33.017 "strip_size_kb": 0, 00:14:33.017 "state": "online", 00:14:33.017 "raid_level": "raid1", 00:14:33.017 "superblock": true, 00:14:33.017 "num_base_bdevs": 4, 00:14:33.017 "num_base_bdevs_discovered": 3, 00:14:33.017 "num_base_bdevs_operational": 3, 00:14:33.017 "process": { 00:14:33.017 "type": "rebuild", 00:14:33.017 "target": "spare", 00:14:33.017 "progress": { 00:14:33.017 "blocks": 20480, 00:14:33.017 "percent": 32 00:14:33.017 } 00:14:33.017 }, 00:14:33.017 "base_bdevs_list": [ 00:14:33.017 { 00:14:33.017 "name": "spare", 00:14:33.017 "uuid": "1bfec8bc-e5c1-5e7b-b8b1-32c7631a4506", 00:14:33.017 "is_configured": true, 00:14:33.017 "data_offset": 2048, 00:14:33.017 "data_size": 63488 00:14:33.017 }, 00:14:33.017 { 00:14:33.017 "name": null, 00:14:33.017 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:33.017 "is_configured": false, 00:14:33.017 "data_offset": 2048, 00:14:33.017 "data_size": 63488 00:14:33.017 }, 00:14:33.017 { 00:14:33.017 "name": "BaseBdev3", 00:14:33.017 "uuid": "0c9a225c-51cb-52ba-8f5c-eba176eb7ad0", 00:14:33.017 "is_configured": true, 00:14:33.017 "data_offset": 2048, 00:14:33.017 "data_size": 63488 00:14:33.017 }, 00:14:33.017 { 00:14:33.017 "name": "BaseBdev4", 00:14:33.017 "uuid": "8fe4c29c-700e-52c1-a0ff-21d9f1e42fe8", 00:14:33.017 "is_configured": true, 00:14:33.017 "data_offset": 2048, 00:14:33.017 "data_size": 63488 00:14:33.017 } 00:14:33.017 ] 00:14:33.017 }' 00:14:33.017 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:33.017 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:33.017 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:33.017 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:33.017 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:14:33.017 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:33.017 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:33.017 [2024-11-19 15:20:23.265892] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:33.017 [2024-11-19 15:20:23.321730] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:14:33.017 [2024-11-19 15:20:23.321788] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:33.017 [2024-11-19 15:20:23.321807] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:33.017 [2024-11-19 15:20:23.321814] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:14:33.017 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:33.017 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:14:33.017 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:33.017 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:33.017 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:33.017 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:33.017 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:33.017 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:33.017 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:33.017 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:33.017 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:33.017 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:33.017 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:33.017 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:33.017 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:33.278 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:33.278 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:33.278 "name": "raid_bdev1", 00:14:33.278 "uuid": "a426f079-d827-4102-96e5-ce4cc6c6d4c0", 00:14:33.278 "strip_size_kb": 0, 00:14:33.278 "state": "online", 00:14:33.278 "raid_level": "raid1", 00:14:33.278 "superblock": true, 00:14:33.278 "num_base_bdevs": 4, 00:14:33.278 "num_base_bdevs_discovered": 2, 00:14:33.278 "num_base_bdevs_operational": 2, 00:14:33.278 "base_bdevs_list": [ 00:14:33.278 { 00:14:33.278 "name": null, 00:14:33.278 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:33.278 "is_configured": false, 00:14:33.278 "data_offset": 0, 00:14:33.278 "data_size": 63488 00:14:33.278 }, 00:14:33.278 { 00:14:33.278 "name": null, 00:14:33.278 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:33.278 "is_configured": false, 00:14:33.278 "data_offset": 2048, 00:14:33.278 "data_size": 63488 00:14:33.278 }, 00:14:33.278 { 00:14:33.278 "name": "BaseBdev3", 00:14:33.278 "uuid": "0c9a225c-51cb-52ba-8f5c-eba176eb7ad0", 00:14:33.278 "is_configured": true, 00:14:33.278 "data_offset": 2048, 00:14:33.278 "data_size": 63488 00:14:33.278 }, 00:14:33.278 { 00:14:33.278 "name": "BaseBdev4", 00:14:33.278 "uuid": "8fe4c29c-700e-52c1-a0ff-21d9f1e42fe8", 00:14:33.278 "is_configured": true, 00:14:33.278 "data_offset": 2048, 00:14:33.278 "data_size": 63488 00:14:33.278 } 00:14:33.278 ] 00:14:33.278 }' 00:14:33.278 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:33.278 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:33.538 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:33.538 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:33.538 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:33.538 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:33.538 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:33.538 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:33.538 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:33.538 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:33.538 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:33.538 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:33.538 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:33.538 "name": "raid_bdev1", 00:14:33.538 "uuid": "a426f079-d827-4102-96e5-ce4cc6c6d4c0", 00:14:33.538 "strip_size_kb": 0, 00:14:33.538 "state": "online", 00:14:33.538 "raid_level": "raid1", 00:14:33.538 "superblock": true, 00:14:33.538 "num_base_bdevs": 4, 00:14:33.538 "num_base_bdevs_discovered": 2, 00:14:33.538 "num_base_bdevs_operational": 2, 00:14:33.538 "base_bdevs_list": [ 00:14:33.538 { 00:14:33.538 "name": null, 00:14:33.538 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:33.538 "is_configured": false, 00:14:33.538 "data_offset": 0, 00:14:33.538 "data_size": 63488 00:14:33.538 }, 00:14:33.538 { 00:14:33.538 "name": null, 00:14:33.538 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:33.538 "is_configured": false, 00:14:33.538 "data_offset": 2048, 00:14:33.538 "data_size": 63488 00:14:33.538 }, 00:14:33.538 { 00:14:33.538 "name": "BaseBdev3", 00:14:33.538 "uuid": "0c9a225c-51cb-52ba-8f5c-eba176eb7ad0", 00:14:33.538 "is_configured": true, 00:14:33.538 "data_offset": 2048, 00:14:33.538 "data_size": 63488 00:14:33.538 }, 00:14:33.538 { 00:14:33.538 "name": "BaseBdev4", 00:14:33.538 "uuid": "8fe4c29c-700e-52c1-a0ff-21d9f1e42fe8", 00:14:33.538 "is_configured": true, 00:14:33.538 "data_offset": 2048, 00:14:33.538 "data_size": 63488 00:14:33.538 } 00:14:33.538 ] 00:14:33.538 }' 00:14:33.538 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:33.538 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:33.538 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:33.538 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:33.538 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:14:33.538 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:33.538 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:33.538 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:33.538 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:14:33.538 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:33.538 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:33.538 [2024-11-19 15:20:23.837412] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:14:33.538 [2024-11-19 15:20:23.837534] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:33.538 [2024-11-19 15:20:23.837563] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c080 00:14:33.538 [2024-11-19 15:20:23.837572] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:33.538 [2024-11-19 15:20:23.837962] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:33.538 [2024-11-19 15:20:23.837992] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:14:33.538 [2024-11-19 15:20:23.838067] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:14:33.538 [2024-11-19 15:20:23.838080] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:14:33.538 [2024-11-19 15:20:23.838089] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:14:33.538 [2024-11-19 15:20:23.838103] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:14:33.538 BaseBdev1 00:14:33.538 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:33.538 15:20:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@775 -- # sleep 1 00:14:34.921 15:20:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:14:34.921 15:20:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:34.921 15:20:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:34.921 15:20:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:34.921 15:20:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:34.921 15:20:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:34.921 15:20:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:34.921 15:20:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:34.921 15:20:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:34.921 15:20:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:34.921 15:20:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:34.921 15:20:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:34.921 15:20:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:34.921 15:20:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:34.921 15:20:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:34.921 15:20:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:34.921 "name": "raid_bdev1", 00:14:34.921 "uuid": "a426f079-d827-4102-96e5-ce4cc6c6d4c0", 00:14:34.921 "strip_size_kb": 0, 00:14:34.921 "state": "online", 00:14:34.921 "raid_level": "raid1", 00:14:34.921 "superblock": true, 00:14:34.921 "num_base_bdevs": 4, 00:14:34.921 "num_base_bdevs_discovered": 2, 00:14:34.921 "num_base_bdevs_operational": 2, 00:14:34.921 "base_bdevs_list": [ 00:14:34.921 { 00:14:34.921 "name": null, 00:14:34.921 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:34.921 "is_configured": false, 00:14:34.921 "data_offset": 0, 00:14:34.921 "data_size": 63488 00:14:34.921 }, 00:14:34.921 { 00:14:34.921 "name": null, 00:14:34.921 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:34.921 "is_configured": false, 00:14:34.921 "data_offset": 2048, 00:14:34.921 "data_size": 63488 00:14:34.921 }, 00:14:34.921 { 00:14:34.921 "name": "BaseBdev3", 00:14:34.921 "uuid": "0c9a225c-51cb-52ba-8f5c-eba176eb7ad0", 00:14:34.921 "is_configured": true, 00:14:34.921 "data_offset": 2048, 00:14:34.921 "data_size": 63488 00:14:34.921 }, 00:14:34.921 { 00:14:34.921 "name": "BaseBdev4", 00:14:34.921 "uuid": "8fe4c29c-700e-52c1-a0ff-21d9f1e42fe8", 00:14:34.921 "is_configured": true, 00:14:34.921 "data_offset": 2048, 00:14:34.921 "data_size": 63488 00:14:34.921 } 00:14:34.921 ] 00:14:34.921 }' 00:14:34.921 15:20:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:34.921 15:20:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:35.181 15:20:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:35.181 15:20:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:35.181 15:20:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:35.181 15:20:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:35.181 15:20:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:35.181 15:20:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:35.181 15:20:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:35.181 15:20:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:35.181 15:20:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:35.181 15:20:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:35.181 15:20:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:35.181 "name": "raid_bdev1", 00:14:35.181 "uuid": "a426f079-d827-4102-96e5-ce4cc6c6d4c0", 00:14:35.181 "strip_size_kb": 0, 00:14:35.181 "state": "online", 00:14:35.181 "raid_level": "raid1", 00:14:35.181 "superblock": true, 00:14:35.181 "num_base_bdevs": 4, 00:14:35.181 "num_base_bdevs_discovered": 2, 00:14:35.181 "num_base_bdevs_operational": 2, 00:14:35.181 "base_bdevs_list": [ 00:14:35.181 { 00:14:35.181 "name": null, 00:14:35.181 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:35.181 "is_configured": false, 00:14:35.181 "data_offset": 0, 00:14:35.181 "data_size": 63488 00:14:35.181 }, 00:14:35.181 { 00:14:35.181 "name": null, 00:14:35.181 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:35.181 "is_configured": false, 00:14:35.181 "data_offset": 2048, 00:14:35.181 "data_size": 63488 00:14:35.181 }, 00:14:35.181 { 00:14:35.181 "name": "BaseBdev3", 00:14:35.181 "uuid": "0c9a225c-51cb-52ba-8f5c-eba176eb7ad0", 00:14:35.181 "is_configured": true, 00:14:35.181 "data_offset": 2048, 00:14:35.181 "data_size": 63488 00:14:35.181 }, 00:14:35.181 { 00:14:35.181 "name": "BaseBdev4", 00:14:35.181 "uuid": "8fe4c29c-700e-52c1-a0ff-21d9f1e42fe8", 00:14:35.181 "is_configured": true, 00:14:35.181 "data_offset": 2048, 00:14:35.181 "data_size": 63488 00:14:35.181 } 00:14:35.181 ] 00:14:35.181 }' 00:14:35.181 15:20:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:35.181 15:20:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:35.181 15:20:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:35.181 15:20:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:35.181 15:20:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:14:35.181 15:20:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@652 -- # local es=0 00:14:35.181 15:20:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:14:35.181 15:20:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:14:35.181 15:20:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:14:35.181 15:20:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:14:35.181 15:20:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:14:35.181 15:20:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:14:35.181 15:20:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:35.181 15:20:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:35.181 [2024-11-19 15:20:25.435259] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:35.181 [2024-11-19 15:20:25.435446] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:14:35.181 [2024-11-19 15:20:25.435465] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:14:35.181 request: 00:14:35.181 { 00:14:35.181 "base_bdev": "BaseBdev1", 00:14:35.181 "raid_bdev": "raid_bdev1", 00:14:35.181 "method": "bdev_raid_add_base_bdev", 00:14:35.181 "req_id": 1 00:14:35.181 } 00:14:35.182 Got JSON-RPC error response 00:14:35.182 response: 00:14:35.182 { 00:14:35.182 "code": -22, 00:14:35.182 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:14:35.182 } 00:14:35.182 15:20:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:14:35.182 15:20:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@655 -- # es=1 00:14:35.182 15:20:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:14:35.182 15:20:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:14:35.182 15:20:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:14:35.182 15:20:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@779 -- # sleep 1 00:14:36.119 15:20:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:14:36.119 15:20:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:36.119 15:20:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:36.119 15:20:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:14:36.119 15:20:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:14:36.119 15:20:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:36.119 15:20:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:36.119 15:20:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:36.119 15:20:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:36.119 15:20:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:36.379 15:20:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:36.379 15:20:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:36.379 15:20:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.379 15:20:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:36.379 15:20:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:36.379 15:20:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:36.379 "name": "raid_bdev1", 00:14:36.379 "uuid": "a426f079-d827-4102-96e5-ce4cc6c6d4c0", 00:14:36.379 "strip_size_kb": 0, 00:14:36.379 "state": "online", 00:14:36.379 "raid_level": "raid1", 00:14:36.379 "superblock": true, 00:14:36.379 "num_base_bdevs": 4, 00:14:36.379 "num_base_bdevs_discovered": 2, 00:14:36.379 "num_base_bdevs_operational": 2, 00:14:36.379 "base_bdevs_list": [ 00:14:36.379 { 00:14:36.379 "name": null, 00:14:36.379 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:36.379 "is_configured": false, 00:14:36.379 "data_offset": 0, 00:14:36.379 "data_size": 63488 00:14:36.379 }, 00:14:36.379 { 00:14:36.379 "name": null, 00:14:36.379 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:36.379 "is_configured": false, 00:14:36.379 "data_offset": 2048, 00:14:36.379 "data_size": 63488 00:14:36.379 }, 00:14:36.379 { 00:14:36.379 "name": "BaseBdev3", 00:14:36.379 "uuid": "0c9a225c-51cb-52ba-8f5c-eba176eb7ad0", 00:14:36.379 "is_configured": true, 00:14:36.379 "data_offset": 2048, 00:14:36.379 "data_size": 63488 00:14:36.379 }, 00:14:36.379 { 00:14:36.379 "name": "BaseBdev4", 00:14:36.379 "uuid": "8fe4c29c-700e-52c1-a0ff-21d9f1e42fe8", 00:14:36.379 "is_configured": true, 00:14:36.379 "data_offset": 2048, 00:14:36.379 "data_size": 63488 00:14:36.379 } 00:14:36.379 ] 00:14:36.379 }' 00:14:36.379 15:20:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:36.379 15:20:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:36.639 15:20:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:36.639 15:20:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:36.639 15:20:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:36.639 15:20:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:36.639 15:20:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:36.639 15:20:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:36.639 15:20:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:36.639 15:20:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.639 15:20:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:36.639 15:20:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:36.639 15:20:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:36.639 "name": "raid_bdev1", 00:14:36.639 "uuid": "a426f079-d827-4102-96e5-ce4cc6c6d4c0", 00:14:36.639 "strip_size_kb": 0, 00:14:36.639 "state": "online", 00:14:36.639 "raid_level": "raid1", 00:14:36.639 "superblock": true, 00:14:36.639 "num_base_bdevs": 4, 00:14:36.639 "num_base_bdevs_discovered": 2, 00:14:36.639 "num_base_bdevs_operational": 2, 00:14:36.639 "base_bdevs_list": [ 00:14:36.639 { 00:14:36.639 "name": null, 00:14:36.639 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:36.639 "is_configured": false, 00:14:36.639 "data_offset": 0, 00:14:36.639 "data_size": 63488 00:14:36.639 }, 00:14:36.639 { 00:14:36.639 "name": null, 00:14:36.639 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:36.639 "is_configured": false, 00:14:36.639 "data_offset": 2048, 00:14:36.639 "data_size": 63488 00:14:36.640 }, 00:14:36.640 { 00:14:36.640 "name": "BaseBdev3", 00:14:36.640 "uuid": "0c9a225c-51cb-52ba-8f5c-eba176eb7ad0", 00:14:36.640 "is_configured": true, 00:14:36.640 "data_offset": 2048, 00:14:36.640 "data_size": 63488 00:14:36.640 }, 00:14:36.640 { 00:14:36.640 "name": "BaseBdev4", 00:14:36.640 "uuid": "8fe4c29c-700e-52c1-a0ff-21d9f1e42fe8", 00:14:36.640 "is_configured": true, 00:14:36.640 "data_offset": 2048, 00:14:36.640 "data_size": 63488 00:14:36.640 } 00:14:36.640 ] 00:14:36.640 }' 00:14:36.640 15:20:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:36.640 15:20:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:36.640 15:20:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:36.900 15:20:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:36.900 15:20:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@784 -- # killprocess 89761 00:14:36.900 15:20:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@954 -- # '[' -z 89761 ']' 00:14:36.900 15:20:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@958 -- # kill -0 89761 00:14:36.900 15:20:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@959 -- # uname 00:14:36.900 15:20:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:14:36.900 15:20:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 89761 00:14:36.900 15:20:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:14:36.900 15:20:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:14:36.900 15:20:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@972 -- # echo 'killing process with pid 89761' 00:14:36.900 killing process with pid 89761 00:14:36.900 15:20:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@973 -- # kill 89761 00:14:36.900 Received shutdown signal, test time was about 17.449889 seconds 00:14:36.900 00:14:36.900 Latency(us) 00:14:36.900 [2024-11-19T15:20:27.239Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:14:36.900 [2024-11-19T15:20:27.239Z] =================================================================================================================== 00:14:36.900 [2024-11-19T15:20:27.239Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:14:36.900 [2024-11-19 15:20:27.054500] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:36.900 [2024-11-19 15:20:27.054650] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:36.900 15:20:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@978 -- # wait 89761 00:14:36.900 [2024-11-19 15:20:27.054716] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:36.900 [2024-11-19 15:20:27.054728] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:14:36.900 [2024-11-19 15:20:27.101116] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:37.160 15:20:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@786 -- # return 0 00:14:37.160 00:14:37.160 real 0m19.417s 00:14:37.160 user 0m25.870s 00:14:37.160 sys 0m2.625s 00:14:37.160 15:20:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1130 -- # xtrace_disable 00:14:37.160 15:20:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:14:37.160 ************************************ 00:14:37.160 END TEST raid_rebuild_test_sb_io 00:14:37.160 ************************************ 00:14:37.160 15:20:27 bdev_raid -- bdev/bdev_raid.sh@985 -- # for n in {3..4} 00:14:37.161 15:20:27 bdev_raid -- bdev/bdev_raid.sh@986 -- # run_test raid5f_state_function_test raid_state_function_test raid5f 3 false 00:14:37.161 15:20:27 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:14:37.161 15:20:27 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:14:37.161 15:20:27 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:37.161 ************************************ 00:14:37.161 START TEST raid5f_state_function_test 00:14:37.161 ************************************ 00:14:37.161 15:20:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid5f 3 false 00:14:37.161 15:20:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:14:37.161 15:20:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:14:37.161 15:20:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:14:37.161 15:20:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:14:37.161 15:20:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:14:37.161 15:20:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:37.161 15:20:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:14:37.161 15:20:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:37.161 15:20:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:37.161 15:20:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:14:37.161 15:20:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:37.161 15:20:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:37.161 15:20:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:14:37.161 15:20:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:37.161 15:20:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:37.161 15:20:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:14:37.161 15:20:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:14:37.161 15:20:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:14:37.161 15:20:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:14:37.161 15:20:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:14:37.161 15:20:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:14:37.161 15:20:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:14:37.161 15:20:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:14:37.161 15:20:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:14:37.161 15:20:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:14:37.161 15:20:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:14:37.161 15:20:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=90462 00:14:37.161 15:20:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:14:37.161 15:20:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 90462' 00:14:37.161 Process raid pid: 90462 00:14:37.161 15:20:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 90462 00:14:37.161 15:20:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 90462 ']' 00:14:37.161 15:20:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:37.161 15:20:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:14:37.161 15:20:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:37.161 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:37.161 15:20:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:14:37.161 15:20:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:37.161 [2024-11-19 15:20:27.481757] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:14:37.161 [2024-11-19 15:20:27.481920] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:14:37.421 [2024-11-19 15:20:27.639188] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:37.421 [2024-11-19 15:20:27.665553] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:14:37.421 [2024-11-19 15:20:27.709705] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:37.421 [2024-11-19 15:20:27.709738] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:37.991 15:20:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:14:37.991 15:20:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:14:37.991 15:20:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:14:37.991 15:20:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:37.991 15:20:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:37.991 [2024-11-19 15:20:28.295875] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:37.991 [2024-11-19 15:20:28.295929] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:37.991 [2024-11-19 15:20:28.295939] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:37.991 [2024-11-19 15:20:28.296056] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:37.991 [2024-11-19 15:20:28.296063] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:37.991 [2024-11-19 15:20:28.296075] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:37.991 15:20:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:37.991 15:20:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:37.991 15:20:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:37.991 15:20:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:37.991 15:20:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:37.991 15:20:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:37.992 15:20:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:37.992 15:20:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:37.992 15:20:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:37.992 15:20:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:37.992 15:20:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:37.992 15:20:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:37.992 15:20:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:37.992 15:20:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:37.992 15:20:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:37.992 15:20:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:38.252 15:20:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:38.252 "name": "Existed_Raid", 00:14:38.252 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:38.252 "strip_size_kb": 64, 00:14:38.252 "state": "configuring", 00:14:38.252 "raid_level": "raid5f", 00:14:38.252 "superblock": false, 00:14:38.252 "num_base_bdevs": 3, 00:14:38.252 "num_base_bdevs_discovered": 0, 00:14:38.252 "num_base_bdevs_operational": 3, 00:14:38.252 "base_bdevs_list": [ 00:14:38.252 { 00:14:38.252 "name": "BaseBdev1", 00:14:38.252 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:38.252 "is_configured": false, 00:14:38.252 "data_offset": 0, 00:14:38.252 "data_size": 0 00:14:38.252 }, 00:14:38.252 { 00:14:38.252 "name": "BaseBdev2", 00:14:38.252 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:38.252 "is_configured": false, 00:14:38.252 "data_offset": 0, 00:14:38.252 "data_size": 0 00:14:38.252 }, 00:14:38.252 { 00:14:38.252 "name": "BaseBdev3", 00:14:38.252 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:38.252 "is_configured": false, 00:14:38.252 "data_offset": 0, 00:14:38.252 "data_size": 0 00:14:38.252 } 00:14:38.252 ] 00:14:38.252 }' 00:14:38.252 15:20:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:38.252 15:20:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:38.512 15:20:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:38.512 15:20:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:38.512 15:20:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:38.512 [2024-11-19 15:20:28.758964] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:38.512 [2024-11-19 15:20:28.759060] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:14:38.512 15:20:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:38.512 15:20:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:14:38.512 15:20:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:38.512 15:20:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:38.512 [2024-11-19 15:20:28.766995] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:38.512 [2024-11-19 15:20:28.767072] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:38.512 [2024-11-19 15:20:28.767102] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:38.512 [2024-11-19 15:20:28.767125] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:38.512 [2024-11-19 15:20:28.767149] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:38.512 [2024-11-19 15:20:28.767171] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:38.512 15:20:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:38.512 15:20:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:14:38.512 15:20:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:38.512 15:20:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:38.512 [2024-11-19 15:20:28.784018] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:38.512 BaseBdev1 00:14:38.512 15:20:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:38.512 15:20:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:14:38.512 15:20:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:14:38.512 15:20:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:38.512 15:20:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:14:38.512 15:20:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:38.512 15:20:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:38.512 15:20:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:38.512 15:20:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:38.512 15:20:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:38.512 15:20:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:38.512 15:20:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:14:38.512 15:20:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:38.512 15:20:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:38.512 [ 00:14:38.512 { 00:14:38.512 "name": "BaseBdev1", 00:14:38.512 "aliases": [ 00:14:38.512 "02fcf116-3d48-4432-b65b-e5a2bc2e8f76" 00:14:38.512 ], 00:14:38.512 "product_name": "Malloc disk", 00:14:38.512 "block_size": 512, 00:14:38.512 "num_blocks": 65536, 00:14:38.512 "uuid": "02fcf116-3d48-4432-b65b-e5a2bc2e8f76", 00:14:38.512 "assigned_rate_limits": { 00:14:38.512 "rw_ios_per_sec": 0, 00:14:38.512 "rw_mbytes_per_sec": 0, 00:14:38.512 "r_mbytes_per_sec": 0, 00:14:38.512 "w_mbytes_per_sec": 0 00:14:38.512 }, 00:14:38.512 "claimed": true, 00:14:38.512 "claim_type": "exclusive_write", 00:14:38.512 "zoned": false, 00:14:38.512 "supported_io_types": { 00:14:38.512 "read": true, 00:14:38.512 "write": true, 00:14:38.512 "unmap": true, 00:14:38.512 "flush": true, 00:14:38.512 "reset": true, 00:14:38.512 "nvme_admin": false, 00:14:38.512 "nvme_io": false, 00:14:38.512 "nvme_io_md": false, 00:14:38.512 "write_zeroes": true, 00:14:38.512 "zcopy": true, 00:14:38.512 "get_zone_info": false, 00:14:38.513 "zone_management": false, 00:14:38.513 "zone_append": false, 00:14:38.513 "compare": false, 00:14:38.513 "compare_and_write": false, 00:14:38.513 "abort": true, 00:14:38.513 "seek_hole": false, 00:14:38.513 "seek_data": false, 00:14:38.513 "copy": true, 00:14:38.513 "nvme_iov_md": false 00:14:38.513 }, 00:14:38.513 "memory_domains": [ 00:14:38.513 { 00:14:38.513 "dma_device_id": "system", 00:14:38.513 "dma_device_type": 1 00:14:38.513 }, 00:14:38.513 { 00:14:38.513 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:38.513 "dma_device_type": 2 00:14:38.513 } 00:14:38.513 ], 00:14:38.513 "driver_specific": {} 00:14:38.513 } 00:14:38.513 ] 00:14:38.513 15:20:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:38.513 15:20:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:14:38.513 15:20:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:38.513 15:20:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:38.513 15:20:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:38.513 15:20:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:38.513 15:20:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:38.513 15:20:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:38.513 15:20:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:38.513 15:20:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:38.513 15:20:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:38.513 15:20:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:38.513 15:20:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:38.513 15:20:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:38.513 15:20:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:38.513 15:20:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:38.513 15:20:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:38.773 15:20:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:38.773 "name": "Existed_Raid", 00:14:38.773 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:38.773 "strip_size_kb": 64, 00:14:38.773 "state": "configuring", 00:14:38.773 "raid_level": "raid5f", 00:14:38.773 "superblock": false, 00:14:38.773 "num_base_bdevs": 3, 00:14:38.773 "num_base_bdevs_discovered": 1, 00:14:38.773 "num_base_bdevs_operational": 3, 00:14:38.773 "base_bdevs_list": [ 00:14:38.773 { 00:14:38.773 "name": "BaseBdev1", 00:14:38.773 "uuid": "02fcf116-3d48-4432-b65b-e5a2bc2e8f76", 00:14:38.773 "is_configured": true, 00:14:38.773 "data_offset": 0, 00:14:38.773 "data_size": 65536 00:14:38.773 }, 00:14:38.773 { 00:14:38.773 "name": "BaseBdev2", 00:14:38.773 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:38.773 "is_configured": false, 00:14:38.773 "data_offset": 0, 00:14:38.773 "data_size": 0 00:14:38.773 }, 00:14:38.773 { 00:14:38.773 "name": "BaseBdev3", 00:14:38.773 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:38.773 "is_configured": false, 00:14:38.773 "data_offset": 0, 00:14:38.773 "data_size": 0 00:14:38.773 } 00:14:38.773 ] 00:14:38.773 }' 00:14:38.773 15:20:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:38.773 15:20:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:39.033 15:20:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:39.033 15:20:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:39.033 15:20:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:39.033 [2024-11-19 15:20:29.239267] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:39.033 [2024-11-19 15:20:29.239307] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:14:39.033 15:20:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:39.033 15:20:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:14:39.033 15:20:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:39.033 15:20:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:39.033 [2024-11-19 15:20:29.251288] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:39.033 [2024-11-19 15:20:29.253134] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:39.033 [2024-11-19 15:20:29.253177] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:39.033 [2024-11-19 15:20:29.253186] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:39.033 [2024-11-19 15:20:29.253197] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:39.033 15:20:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:39.033 15:20:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:14:39.033 15:20:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:39.033 15:20:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:39.033 15:20:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:39.033 15:20:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:39.033 15:20:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:39.033 15:20:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:39.033 15:20:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:39.033 15:20:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:39.033 15:20:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:39.033 15:20:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:39.033 15:20:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:39.033 15:20:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:39.033 15:20:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:39.033 15:20:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:39.033 15:20:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:39.033 15:20:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:39.033 15:20:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:39.033 "name": "Existed_Raid", 00:14:39.033 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:39.033 "strip_size_kb": 64, 00:14:39.033 "state": "configuring", 00:14:39.033 "raid_level": "raid5f", 00:14:39.033 "superblock": false, 00:14:39.033 "num_base_bdevs": 3, 00:14:39.033 "num_base_bdevs_discovered": 1, 00:14:39.033 "num_base_bdevs_operational": 3, 00:14:39.033 "base_bdevs_list": [ 00:14:39.033 { 00:14:39.033 "name": "BaseBdev1", 00:14:39.033 "uuid": "02fcf116-3d48-4432-b65b-e5a2bc2e8f76", 00:14:39.033 "is_configured": true, 00:14:39.033 "data_offset": 0, 00:14:39.033 "data_size": 65536 00:14:39.033 }, 00:14:39.033 { 00:14:39.033 "name": "BaseBdev2", 00:14:39.033 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:39.033 "is_configured": false, 00:14:39.033 "data_offset": 0, 00:14:39.033 "data_size": 0 00:14:39.033 }, 00:14:39.033 { 00:14:39.033 "name": "BaseBdev3", 00:14:39.033 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:39.034 "is_configured": false, 00:14:39.034 "data_offset": 0, 00:14:39.034 "data_size": 0 00:14:39.034 } 00:14:39.034 ] 00:14:39.034 }' 00:14:39.034 15:20:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:39.034 15:20:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:39.603 15:20:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:14:39.603 15:20:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:39.603 15:20:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:39.603 [2024-11-19 15:20:29.657751] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:39.603 BaseBdev2 00:14:39.603 15:20:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:39.603 15:20:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:14:39.603 15:20:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:14:39.603 15:20:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:39.603 15:20:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:14:39.603 15:20:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:39.603 15:20:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:39.603 15:20:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:39.603 15:20:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:39.603 15:20:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:39.603 15:20:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:39.603 15:20:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:14:39.603 15:20:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:39.603 15:20:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:39.603 [ 00:14:39.603 { 00:14:39.603 "name": "BaseBdev2", 00:14:39.603 "aliases": [ 00:14:39.603 "9a3f0fbb-c522-4d40-9065-10adb7f8dbff" 00:14:39.603 ], 00:14:39.603 "product_name": "Malloc disk", 00:14:39.603 "block_size": 512, 00:14:39.603 "num_blocks": 65536, 00:14:39.603 "uuid": "9a3f0fbb-c522-4d40-9065-10adb7f8dbff", 00:14:39.603 "assigned_rate_limits": { 00:14:39.603 "rw_ios_per_sec": 0, 00:14:39.603 "rw_mbytes_per_sec": 0, 00:14:39.603 "r_mbytes_per_sec": 0, 00:14:39.603 "w_mbytes_per_sec": 0 00:14:39.603 }, 00:14:39.603 "claimed": true, 00:14:39.603 "claim_type": "exclusive_write", 00:14:39.603 "zoned": false, 00:14:39.604 "supported_io_types": { 00:14:39.604 "read": true, 00:14:39.604 "write": true, 00:14:39.604 "unmap": true, 00:14:39.604 "flush": true, 00:14:39.604 "reset": true, 00:14:39.604 "nvme_admin": false, 00:14:39.604 "nvme_io": false, 00:14:39.604 "nvme_io_md": false, 00:14:39.604 "write_zeroes": true, 00:14:39.604 "zcopy": true, 00:14:39.604 "get_zone_info": false, 00:14:39.604 "zone_management": false, 00:14:39.604 "zone_append": false, 00:14:39.604 "compare": false, 00:14:39.604 "compare_and_write": false, 00:14:39.604 "abort": true, 00:14:39.604 "seek_hole": false, 00:14:39.604 "seek_data": false, 00:14:39.604 "copy": true, 00:14:39.604 "nvme_iov_md": false 00:14:39.604 }, 00:14:39.604 "memory_domains": [ 00:14:39.604 { 00:14:39.604 "dma_device_id": "system", 00:14:39.604 "dma_device_type": 1 00:14:39.604 }, 00:14:39.604 { 00:14:39.604 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:39.604 "dma_device_type": 2 00:14:39.604 } 00:14:39.604 ], 00:14:39.604 "driver_specific": {} 00:14:39.604 } 00:14:39.604 ] 00:14:39.604 15:20:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:39.604 15:20:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:14:39.604 15:20:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:39.604 15:20:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:39.604 15:20:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:39.604 15:20:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:39.604 15:20:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:39.604 15:20:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:39.604 15:20:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:39.604 15:20:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:39.604 15:20:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:39.604 15:20:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:39.604 15:20:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:39.604 15:20:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:39.604 15:20:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:39.604 15:20:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:39.604 15:20:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:39.604 15:20:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:39.604 15:20:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:39.604 15:20:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:39.604 "name": "Existed_Raid", 00:14:39.604 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:39.604 "strip_size_kb": 64, 00:14:39.604 "state": "configuring", 00:14:39.604 "raid_level": "raid5f", 00:14:39.604 "superblock": false, 00:14:39.604 "num_base_bdevs": 3, 00:14:39.604 "num_base_bdevs_discovered": 2, 00:14:39.604 "num_base_bdevs_operational": 3, 00:14:39.604 "base_bdevs_list": [ 00:14:39.604 { 00:14:39.604 "name": "BaseBdev1", 00:14:39.604 "uuid": "02fcf116-3d48-4432-b65b-e5a2bc2e8f76", 00:14:39.604 "is_configured": true, 00:14:39.604 "data_offset": 0, 00:14:39.604 "data_size": 65536 00:14:39.604 }, 00:14:39.604 { 00:14:39.604 "name": "BaseBdev2", 00:14:39.604 "uuid": "9a3f0fbb-c522-4d40-9065-10adb7f8dbff", 00:14:39.604 "is_configured": true, 00:14:39.604 "data_offset": 0, 00:14:39.604 "data_size": 65536 00:14:39.604 }, 00:14:39.604 { 00:14:39.604 "name": "BaseBdev3", 00:14:39.604 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:39.604 "is_configured": false, 00:14:39.604 "data_offset": 0, 00:14:39.604 "data_size": 0 00:14:39.604 } 00:14:39.604 ] 00:14:39.604 }' 00:14:39.604 15:20:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:39.604 15:20:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:39.864 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:14:39.864 15:20:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:39.864 15:20:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:39.864 [2024-11-19 15:20:30.174859] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:39.864 [2024-11-19 15:20:30.175067] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:14:39.864 [2024-11-19 15:20:30.175151] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:14:39.864 [2024-11-19 15:20:30.176123] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:14:39.864 [2024-11-19 15:20:30.177821] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:14:39.864 [2024-11-19 15:20:30.177882] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:14:39.864 [2024-11-19 15:20:30.178587] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:39.864 BaseBdev3 00:14:39.864 15:20:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:39.864 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:14:39.864 15:20:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:14:39.864 15:20:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:39.864 15:20:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:14:39.864 15:20:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:39.864 15:20:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:39.864 15:20:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:39.864 15:20:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:39.864 15:20:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:39.864 15:20:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:39.864 15:20:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:14:39.864 15:20:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:39.864 15:20:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:40.124 [ 00:14:40.124 { 00:14:40.124 "name": "BaseBdev3", 00:14:40.124 "aliases": [ 00:14:40.124 "2de4cef4-1dbc-4ce3-956b-65953004219c" 00:14:40.124 ], 00:14:40.124 "product_name": "Malloc disk", 00:14:40.124 "block_size": 512, 00:14:40.124 "num_blocks": 65536, 00:14:40.124 "uuid": "2de4cef4-1dbc-4ce3-956b-65953004219c", 00:14:40.124 "assigned_rate_limits": { 00:14:40.124 "rw_ios_per_sec": 0, 00:14:40.124 "rw_mbytes_per_sec": 0, 00:14:40.124 "r_mbytes_per_sec": 0, 00:14:40.124 "w_mbytes_per_sec": 0 00:14:40.124 }, 00:14:40.124 "claimed": true, 00:14:40.124 "claim_type": "exclusive_write", 00:14:40.124 "zoned": false, 00:14:40.124 "supported_io_types": { 00:14:40.124 "read": true, 00:14:40.124 "write": true, 00:14:40.124 "unmap": true, 00:14:40.124 "flush": true, 00:14:40.124 "reset": true, 00:14:40.124 "nvme_admin": false, 00:14:40.124 "nvme_io": false, 00:14:40.124 "nvme_io_md": false, 00:14:40.124 "write_zeroes": true, 00:14:40.124 "zcopy": true, 00:14:40.124 "get_zone_info": false, 00:14:40.124 "zone_management": false, 00:14:40.124 "zone_append": false, 00:14:40.124 "compare": false, 00:14:40.124 "compare_and_write": false, 00:14:40.125 "abort": true, 00:14:40.125 "seek_hole": false, 00:14:40.125 "seek_data": false, 00:14:40.125 "copy": true, 00:14:40.125 "nvme_iov_md": false 00:14:40.125 }, 00:14:40.125 "memory_domains": [ 00:14:40.125 { 00:14:40.125 "dma_device_id": "system", 00:14:40.125 "dma_device_type": 1 00:14:40.125 }, 00:14:40.125 { 00:14:40.125 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:40.125 "dma_device_type": 2 00:14:40.125 } 00:14:40.125 ], 00:14:40.125 "driver_specific": {} 00:14:40.125 } 00:14:40.125 ] 00:14:40.125 15:20:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:40.125 15:20:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:14:40.125 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:40.125 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:40.125 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:14:40.125 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:40.125 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:40.125 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:40.125 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:40.125 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:40.125 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:40.125 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:40.125 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:40.125 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:40.125 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:40.125 15:20:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:40.125 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:40.125 15:20:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:40.125 15:20:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:40.125 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:40.125 "name": "Existed_Raid", 00:14:40.125 "uuid": "2c8071d0-7f84-4bfa-83e4-54ca2b9304a8", 00:14:40.125 "strip_size_kb": 64, 00:14:40.125 "state": "online", 00:14:40.125 "raid_level": "raid5f", 00:14:40.125 "superblock": false, 00:14:40.125 "num_base_bdevs": 3, 00:14:40.125 "num_base_bdevs_discovered": 3, 00:14:40.125 "num_base_bdevs_operational": 3, 00:14:40.125 "base_bdevs_list": [ 00:14:40.125 { 00:14:40.125 "name": "BaseBdev1", 00:14:40.125 "uuid": "02fcf116-3d48-4432-b65b-e5a2bc2e8f76", 00:14:40.125 "is_configured": true, 00:14:40.125 "data_offset": 0, 00:14:40.125 "data_size": 65536 00:14:40.125 }, 00:14:40.125 { 00:14:40.125 "name": "BaseBdev2", 00:14:40.125 "uuid": "9a3f0fbb-c522-4d40-9065-10adb7f8dbff", 00:14:40.125 "is_configured": true, 00:14:40.125 "data_offset": 0, 00:14:40.125 "data_size": 65536 00:14:40.125 }, 00:14:40.125 { 00:14:40.125 "name": "BaseBdev3", 00:14:40.125 "uuid": "2de4cef4-1dbc-4ce3-956b-65953004219c", 00:14:40.125 "is_configured": true, 00:14:40.125 "data_offset": 0, 00:14:40.125 "data_size": 65536 00:14:40.125 } 00:14:40.125 ] 00:14:40.125 }' 00:14:40.125 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:40.125 15:20:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:40.385 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:14:40.385 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:14:40.385 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:40.385 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:40.385 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:14:40.385 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:40.385 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:40.385 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:14:40.385 15:20:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:40.385 15:20:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:40.385 [2024-11-19 15:20:30.688691] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:40.385 15:20:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:40.645 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:40.645 "name": "Existed_Raid", 00:14:40.645 "aliases": [ 00:14:40.645 "2c8071d0-7f84-4bfa-83e4-54ca2b9304a8" 00:14:40.645 ], 00:14:40.645 "product_name": "Raid Volume", 00:14:40.645 "block_size": 512, 00:14:40.645 "num_blocks": 131072, 00:14:40.645 "uuid": "2c8071d0-7f84-4bfa-83e4-54ca2b9304a8", 00:14:40.645 "assigned_rate_limits": { 00:14:40.645 "rw_ios_per_sec": 0, 00:14:40.645 "rw_mbytes_per_sec": 0, 00:14:40.645 "r_mbytes_per_sec": 0, 00:14:40.645 "w_mbytes_per_sec": 0 00:14:40.645 }, 00:14:40.645 "claimed": false, 00:14:40.645 "zoned": false, 00:14:40.645 "supported_io_types": { 00:14:40.645 "read": true, 00:14:40.645 "write": true, 00:14:40.645 "unmap": false, 00:14:40.645 "flush": false, 00:14:40.645 "reset": true, 00:14:40.645 "nvme_admin": false, 00:14:40.645 "nvme_io": false, 00:14:40.645 "nvme_io_md": false, 00:14:40.645 "write_zeroes": true, 00:14:40.645 "zcopy": false, 00:14:40.645 "get_zone_info": false, 00:14:40.645 "zone_management": false, 00:14:40.645 "zone_append": false, 00:14:40.645 "compare": false, 00:14:40.645 "compare_and_write": false, 00:14:40.645 "abort": false, 00:14:40.645 "seek_hole": false, 00:14:40.645 "seek_data": false, 00:14:40.645 "copy": false, 00:14:40.645 "nvme_iov_md": false 00:14:40.645 }, 00:14:40.645 "driver_specific": { 00:14:40.645 "raid": { 00:14:40.645 "uuid": "2c8071d0-7f84-4bfa-83e4-54ca2b9304a8", 00:14:40.645 "strip_size_kb": 64, 00:14:40.645 "state": "online", 00:14:40.645 "raid_level": "raid5f", 00:14:40.645 "superblock": false, 00:14:40.645 "num_base_bdevs": 3, 00:14:40.645 "num_base_bdevs_discovered": 3, 00:14:40.645 "num_base_bdevs_operational": 3, 00:14:40.645 "base_bdevs_list": [ 00:14:40.645 { 00:14:40.645 "name": "BaseBdev1", 00:14:40.645 "uuid": "02fcf116-3d48-4432-b65b-e5a2bc2e8f76", 00:14:40.645 "is_configured": true, 00:14:40.645 "data_offset": 0, 00:14:40.645 "data_size": 65536 00:14:40.645 }, 00:14:40.645 { 00:14:40.645 "name": "BaseBdev2", 00:14:40.645 "uuid": "9a3f0fbb-c522-4d40-9065-10adb7f8dbff", 00:14:40.645 "is_configured": true, 00:14:40.645 "data_offset": 0, 00:14:40.645 "data_size": 65536 00:14:40.645 }, 00:14:40.645 { 00:14:40.645 "name": "BaseBdev3", 00:14:40.645 "uuid": "2de4cef4-1dbc-4ce3-956b-65953004219c", 00:14:40.645 "is_configured": true, 00:14:40.645 "data_offset": 0, 00:14:40.645 "data_size": 65536 00:14:40.645 } 00:14:40.645 ] 00:14:40.646 } 00:14:40.646 } 00:14:40.646 }' 00:14:40.646 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:40.646 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:14:40.646 BaseBdev2 00:14:40.646 BaseBdev3' 00:14:40.646 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:40.646 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:40.646 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:40.646 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:14:40.646 15:20:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:40.646 15:20:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:40.646 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:40.646 15:20:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:40.646 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:40.646 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:40.646 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:40.646 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:14:40.646 15:20:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:40.646 15:20:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:40.646 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:40.646 15:20:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:40.646 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:40.646 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:40.646 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:40.646 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:40.646 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:14:40.646 15:20:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:40.646 15:20:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:40.646 15:20:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:40.906 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:40.906 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:40.906 15:20:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:14:40.906 15:20:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:40.906 15:20:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:40.906 [2024-11-19 15:20:30.992084] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:40.906 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:40.906 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:14:40.906 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:14:40.906 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:14:40.906 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:14:40.906 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:14:40.906 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 2 00:14:40.906 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:40.906 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:40.906 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:40.906 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:40.906 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:40.906 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:40.906 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:40.906 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:40.906 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:40.906 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:40.906 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:40.906 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:40.906 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:40.906 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:40.906 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:40.906 "name": "Existed_Raid", 00:14:40.906 "uuid": "2c8071d0-7f84-4bfa-83e4-54ca2b9304a8", 00:14:40.906 "strip_size_kb": 64, 00:14:40.906 "state": "online", 00:14:40.906 "raid_level": "raid5f", 00:14:40.906 "superblock": false, 00:14:40.907 "num_base_bdevs": 3, 00:14:40.907 "num_base_bdevs_discovered": 2, 00:14:40.907 "num_base_bdevs_operational": 2, 00:14:40.907 "base_bdevs_list": [ 00:14:40.907 { 00:14:40.907 "name": null, 00:14:40.907 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:40.907 "is_configured": false, 00:14:40.907 "data_offset": 0, 00:14:40.907 "data_size": 65536 00:14:40.907 }, 00:14:40.907 { 00:14:40.907 "name": "BaseBdev2", 00:14:40.907 "uuid": "9a3f0fbb-c522-4d40-9065-10adb7f8dbff", 00:14:40.907 "is_configured": true, 00:14:40.907 "data_offset": 0, 00:14:40.907 "data_size": 65536 00:14:40.907 }, 00:14:40.907 { 00:14:40.907 "name": "BaseBdev3", 00:14:40.907 "uuid": "2de4cef4-1dbc-4ce3-956b-65953004219c", 00:14:40.907 "is_configured": true, 00:14:40.907 "data_offset": 0, 00:14:40.907 "data_size": 65536 00:14:40.907 } 00:14:40.907 ] 00:14:40.907 }' 00:14:40.907 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:40.907 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:41.167 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:14:41.167 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:41.167 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:41.167 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:41.167 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:41.167 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:41.167 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:41.428 [2024-11-19 15:20:31.522521] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:14:41.428 [2024-11-19 15:20:31.522608] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:41.428 [2024-11-19 15:20:31.533764] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:41.428 [2024-11-19 15:20:31.593680] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:14:41.428 [2024-11-19 15:20:31.593721] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:41.428 BaseBdev2 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:41.428 [ 00:14:41.428 { 00:14:41.428 "name": "BaseBdev2", 00:14:41.428 "aliases": [ 00:14:41.428 "367d63c0-9aa4-46b1-a02b-a2171810fa26" 00:14:41.428 ], 00:14:41.428 "product_name": "Malloc disk", 00:14:41.428 "block_size": 512, 00:14:41.428 "num_blocks": 65536, 00:14:41.428 "uuid": "367d63c0-9aa4-46b1-a02b-a2171810fa26", 00:14:41.428 "assigned_rate_limits": { 00:14:41.428 "rw_ios_per_sec": 0, 00:14:41.428 "rw_mbytes_per_sec": 0, 00:14:41.428 "r_mbytes_per_sec": 0, 00:14:41.428 "w_mbytes_per_sec": 0 00:14:41.428 }, 00:14:41.428 "claimed": false, 00:14:41.428 "zoned": false, 00:14:41.428 "supported_io_types": { 00:14:41.428 "read": true, 00:14:41.428 "write": true, 00:14:41.428 "unmap": true, 00:14:41.428 "flush": true, 00:14:41.428 "reset": true, 00:14:41.428 "nvme_admin": false, 00:14:41.428 "nvme_io": false, 00:14:41.428 "nvme_io_md": false, 00:14:41.428 "write_zeroes": true, 00:14:41.428 "zcopy": true, 00:14:41.428 "get_zone_info": false, 00:14:41.428 "zone_management": false, 00:14:41.428 "zone_append": false, 00:14:41.428 "compare": false, 00:14:41.428 "compare_and_write": false, 00:14:41.428 "abort": true, 00:14:41.428 "seek_hole": false, 00:14:41.428 "seek_data": false, 00:14:41.428 "copy": true, 00:14:41.428 "nvme_iov_md": false 00:14:41.428 }, 00:14:41.428 "memory_domains": [ 00:14:41.428 { 00:14:41.428 "dma_device_id": "system", 00:14:41.428 "dma_device_type": 1 00:14:41.428 }, 00:14:41.428 { 00:14:41.428 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:41.428 "dma_device_type": 2 00:14:41.428 } 00:14:41.428 ], 00:14:41.428 "driver_specific": {} 00:14:41.428 } 00:14:41.428 ] 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:41.428 BaseBdev3 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:41.428 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:41.428 [ 00:14:41.428 { 00:14:41.428 "name": "BaseBdev3", 00:14:41.428 "aliases": [ 00:14:41.428 "600f47a6-872d-4e13-89e6-a4a9e2b33425" 00:14:41.428 ], 00:14:41.428 "product_name": "Malloc disk", 00:14:41.428 "block_size": 512, 00:14:41.428 "num_blocks": 65536, 00:14:41.428 "uuid": "600f47a6-872d-4e13-89e6-a4a9e2b33425", 00:14:41.428 "assigned_rate_limits": { 00:14:41.428 "rw_ios_per_sec": 0, 00:14:41.428 "rw_mbytes_per_sec": 0, 00:14:41.428 "r_mbytes_per_sec": 0, 00:14:41.428 "w_mbytes_per_sec": 0 00:14:41.428 }, 00:14:41.428 "claimed": false, 00:14:41.428 "zoned": false, 00:14:41.428 "supported_io_types": { 00:14:41.428 "read": true, 00:14:41.428 "write": true, 00:14:41.428 "unmap": true, 00:14:41.428 "flush": true, 00:14:41.428 "reset": true, 00:14:41.428 "nvme_admin": false, 00:14:41.428 "nvme_io": false, 00:14:41.428 "nvme_io_md": false, 00:14:41.428 "write_zeroes": true, 00:14:41.428 "zcopy": true, 00:14:41.428 "get_zone_info": false, 00:14:41.428 "zone_management": false, 00:14:41.428 "zone_append": false, 00:14:41.429 "compare": false, 00:14:41.429 "compare_and_write": false, 00:14:41.429 "abort": true, 00:14:41.429 "seek_hole": false, 00:14:41.429 "seek_data": false, 00:14:41.429 "copy": true, 00:14:41.429 "nvme_iov_md": false 00:14:41.429 }, 00:14:41.429 "memory_domains": [ 00:14:41.429 { 00:14:41.429 "dma_device_id": "system", 00:14:41.429 "dma_device_type": 1 00:14:41.429 }, 00:14:41.429 { 00:14:41.429 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:41.429 "dma_device_type": 2 00:14:41.429 } 00:14:41.429 ], 00:14:41.429 "driver_specific": {} 00:14:41.429 } 00:14:41.429 ] 00:14:41.429 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:41.429 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:14:41.429 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:41.429 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:41.429 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:14:41.429 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:41.429 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:41.689 [2024-11-19 15:20:31.768096] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:41.689 [2024-11-19 15:20:31.768178] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:41.689 [2024-11-19 15:20:31.768233] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:41.689 [2024-11-19 15:20:31.770018] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:41.689 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:41.689 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:41.689 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:41.689 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:41.689 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:41.689 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:41.689 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:41.689 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:41.689 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:41.689 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:41.689 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:41.689 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:41.689 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:41.689 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:41.689 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:41.689 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:41.689 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:41.689 "name": "Existed_Raid", 00:14:41.689 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:41.689 "strip_size_kb": 64, 00:14:41.689 "state": "configuring", 00:14:41.689 "raid_level": "raid5f", 00:14:41.689 "superblock": false, 00:14:41.689 "num_base_bdevs": 3, 00:14:41.689 "num_base_bdevs_discovered": 2, 00:14:41.689 "num_base_bdevs_operational": 3, 00:14:41.689 "base_bdevs_list": [ 00:14:41.689 { 00:14:41.689 "name": "BaseBdev1", 00:14:41.689 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:41.689 "is_configured": false, 00:14:41.689 "data_offset": 0, 00:14:41.689 "data_size": 0 00:14:41.689 }, 00:14:41.689 { 00:14:41.689 "name": "BaseBdev2", 00:14:41.689 "uuid": "367d63c0-9aa4-46b1-a02b-a2171810fa26", 00:14:41.689 "is_configured": true, 00:14:41.689 "data_offset": 0, 00:14:41.689 "data_size": 65536 00:14:41.689 }, 00:14:41.689 { 00:14:41.689 "name": "BaseBdev3", 00:14:41.689 "uuid": "600f47a6-872d-4e13-89e6-a4a9e2b33425", 00:14:41.689 "is_configured": true, 00:14:41.689 "data_offset": 0, 00:14:41.689 "data_size": 65536 00:14:41.689 } 00:14:41.689 ] 00:14:41.689 }' 00:14:41.689 15:20:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:41.689 15:20:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:41.950 15:20:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:14:41.950 15:20:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:41.950 15:20:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:41.950 [2024-11-19 15:20:32.247558] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:14:41.950 15:20:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:41.950 15:20:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:41.950 15:20:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:41.950 15:20:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:41.950 15:20:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:41.950 15:20:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:41.950 15:20:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:41.950 15:20:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:41.950 15:20:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:41.950 15:20:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:41.950 15:20:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:41.950 15:20:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:41.950 15:20:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:41.950 15:20:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:41.950 15:20:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:41.950 15:20:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:42.210 15:20:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:42.210 "name": "Existed_Raid", 00:14:42.210 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:42.210 "strip_size_kb": 64, 00:14:42.210 "state": "configuring", 00:14:42.210 "raid_level": "raid5f", 00:14:42.210 "superblock": false, 00:14:42.210 "num_base_bdevs": 3, 00:14:42.210 "num_base_bdevs_discovered": 1, 00:14:42.210 "num_base_bdevs_operational": 3, 00:14:42.210 "base_bdevs_list": [ 00:14:42.210 { 00:14:42.210 "name": "BaseBdev1", 00:14:42.210 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:42.210 "is_configured": false, 00:14:42.210 "data_offset": 0, 00:14:42.210 "data_size": 0 00:14:42.210 }, 00:14:42.210 { 00:14:42.210 "name": null, 00:14:42.210 "uuid": "367d63c0-9aa4-46b1-a02b-a2171810fa26", 00:14:42.210 "is_configured": false, 00:14:42.210 "data_offset": 0, 00:14:42.210 "data_size": 65536 00:14:42.210 }, 00:14:42.210 { 00:14:42.210 "name": "BaseBdev3", 00:14:42.210 "uuid": "600f47a6-872d-4e13-89e6-a4a9e2b33425", 00:14:42.210 "is_configured": true, 00:14:42.210 "data_offset": 0, 00:14:42.210 "data_size": 65536 00:14:42.210 } 00:14:42.210 ] 00:14:42.210 }' 00:14:42.210 15:20:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:42.210 15:20:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:42.470 15:20:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:14:42.470 15:20:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:42.470 15:20:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:42.470 15:20:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:42.470 15:20:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:42.470 15:20:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:14:42.470 15:20:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:14:42.470 15:20:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:42.470 15:20:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:42.470 [2024-11-19 15:20:32.725782] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:42.470 BaseBdev1 00:14:42.470 15:20:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:42.470 15:20:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:14:42.470 15:20:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:14:42.470 15:20:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:42.470 15:20:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:14:42.470 15:20:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:42.470 15:20:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:42.470 15:20:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:42.470 15:20:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:42.470 15:20:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:42.470 15:20:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:42.470 15:20:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:14:42.470 15:20:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:42.470 15:20:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:42.470 [ 00:14:42.470 { 00:14:42.470 "name": "BaseBdev1", 00:14:42.470 "aliases": [ 00:14:42.470 "690cf7ac-b179-4ac5-8e53-018b22f92671" 00:14:42.470 ], 00:14:42.470 "product_name": "Malloc disk", 00:14:42.470 "block_size": 512, 00:14:42.470 "num_blocks": 65536, 00:14:42.470 "uuid": "690cf7ac-b179-4ac5-8e53-018b22f92671", 00:14:42.470 "assigned_rate_limits": { 00:14:42.470 "rw_ios_per_sec": 0, 00:14:42.470 "rw_mbytes_per_sec": 0, 00:14:42.470 "r_mbytes_per_sec": 0, 00:14:42.470 "w_mbytes_per_sec": 0 00:14:42.470 }, 00:14:42.470 "claimed": true, 00:14:42.470 "claim_type": "exclusive_write", 00:14:42.470 "zoned": false, 00:14:42.470 "supported_io_types": { 00:14:42.470 "read": true, 00:14:42.470 "write": true, 00:14:42.470 "unmap": true, 00:14:42.470 "flush": true, 00:14:42.470 "reset": true, 00:14:42.470 "nvme_admin": false, 00:14:42.470 "nvme_io": false, 00:14:42.470 "nvme_io_md": false, 00:14:42.470 "write_zeroes": true, 00:14:42.470 "zcopy": true, 00:14:42.470 "get_zone_info": false, 00:14:42.470 "zone_management": false, 00:14:42.470 "zone_append": false, 00:14:42.470 "compare": false, 00:14:42.470 "compare_and_write": false, 00:14:42.470 "abort": true, 00:14:42.470 "seek_hole": false, 00:14:42.470 "seek_data": false, 00:14:42.470 "copy": true, 00:14:42.470 "nvme_iov_md": false 00:14:42.470 }, 00:14:42.470 "memory_domains": [ 00:14:42.470 { 00:14:42.470 "dma_device_id": "system", 00:14:42.470 "dma_device_type": 1 00:14:42.470 }, 00:14:42.470 { 00:14:42.470 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:42.470 "dma_device_type": 2 00:14:42.470 } 00:14:42.470 ], 00:14:42.470 "driver_specific": {} 00:14:42.470 } 00:14:42.470 ] 00:14:42.470 15:20:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:42.470 15:20:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:14:42.470 15:20:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:42.470 15:20:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:42.470 15:20:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:42.470 15:20:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:42.470 15:20:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:42.470 15:20:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:42.470 15:20:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:42.470 15:20:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:42.470 15:20:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:42.470 15:20:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:42.470 15:20:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:42.470 15:20:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:42.470 15:20:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:42.471 15:20:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:42.471 15:20:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:42.731 15:20:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:42.731 "name": "Existed_Raid", 00:14:42.731 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:42.731 "strip_size_kb": 64, 00:14:42.731 "state": "configuring", 00:14:42.731 "raid_level": "raid5f", 00:14:42.731 "superblock": false, 00:14:42.731 "num_base_bdevs": 3, 00:14:42.731 "num_base_bdevs_discovered": 2, 00:14:42.731 "num_base_bdevs_operational": 3, 00:14:42.731 "base_bdevs_list": [ 00:14:42.731 { 00:14:42.731 "name": "BaseBdev1", 00:14:42.731 "uuid": "690cf7ac-b179-4ac5-8e53-018b22f92671", 00:14:42.731 "is_configured": true, 00:14:42.731 "data_offset": 0, 00:14:42.731 "data_size": 65536 00:14:42.731 }, 00:14:42.731 { 00:14:42.731 "name": null, 00:14:42.731 "uuid": "367d63c0-9aa4-46b1-a02b-a2171810fa26", 00:14:42.731 "is_configured": false, 00:14:42.731 "data_offset": 0, 00:14:42.731 "data_size": 65536 00:14:42.731 }, 00:14:42.731 { 00:14:42.731 "name": "BaseBdev3", 00:14:42.731 "uuid": "600f47a6-872d-4e13-89e6-a4a9e2b33425", 00:14:42.731 "is_configured": true, 00:14:42.731 "data_offset": 0, 00:14:42.731 "data_size": 65536 00:14:42.731 } 00:14:42.731 ] 00:14:42.731 }' 00:14:42.731 15:20:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:42.731 15:20:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:43.002 15:20:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:14:43.002 15:20:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:43.002 15:20:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:43.002 15:20:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:43.002 15:20:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:43.002 15:20:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:14:43.002 15:20:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:14:43.002 15:20:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:43.002 15:20:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:43.002 [2024-11-19 15:20:33.288833] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:14:43.002 15:20:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:43.002 15:20:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:43.002 15:20:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:43.002 15:20:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:43.002 15:20:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:43.002 15:20:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:43.002 15:20:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:43.002 15:20:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:43.002 15:20:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:43.002 15:20:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:43.002 15:20:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:43.002 15:20:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:43.002 15:20:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:43.002 15:20:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:43.002 15:20:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:43.002 15:20:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:43.002 15:20:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:43.002 "name": "Existed_Raid", 00:14:43.002 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:43.002 "strip_size_kb": 64, 00:14:43.002 "state": "configuring", 00:14:43.002 "raid_level": "raid5f", 00:14:43.002 "superblock": false, 00:14:43.002 "num_base_bdevs": 3, 00:14:43.002 "num_base_bdevs_discovered": 1, 00:14:43.002 "num_base_bdevs_operational": 3, 00:14:43.002 "base_bdevs_list": [ 00:14:43.002 { 00:14:43.002 "name": "BaseBdev1", 00:14:43.002 "uuid": "690cf7ac-b179-4ac5-8e53-018b22f92671", 00:14:43.002 "is_configured": true, 00:14:43.002 "data_offset": 0, 00:14:43.002 "data_size": 65536 00:14:43.002 }, 00:14:43.002 { 00:14:43.002 "name": null, 00:14:43.002 "uuid": "367d63c0-9aa4-46b1-a02b-a2171810fa26", 00:14:43.002 "is_configured": false, 00:14:43.002 "data_offset": 0, 00:14:43.002 "data_size": 65536 00:14:43.002 }, 00:14:43.002 { 00:14:43.002 "name": null, 00:14:43.002 "uuid": "600f47a6-872d-4e13-89e6-a4a9e2b33425", 00:14:43.002 "is_configured": false, 00:14:43.002 "data_offset": 0, 00:14:43.002 "data_size": 65536 00:14:43.002 } 00:14:43.002 ] 00:14:43.002 }' 00:14:43.002 15:20:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:43.002 15:20:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:43.611 15:20:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:43.611 15:20:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:14:43.611 15:20:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:43.611 15:20:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:43.611 15:20:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:43.611 15:20:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:14:43.611 15:20:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:14:43.611 15:20:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:43.611 15:20:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:43.611 [2024-11-19 15:20:33.764028] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:43.611 15:20:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:43.611 15:20:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:43.611 15:20:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:43.611 15:20:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:43.611 15:20:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:43.611 15:20:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:43.611 15:20:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:43.611 15:20:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:43.611 15:20:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:43.611 15:20:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:43.611 15:20:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:43.611 15:20:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:43.611 15:20:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:43.611 15:20:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:43.611 15:20:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:43.611 15:20:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:43.611 15:20:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:43.611 "name": "Existed_Raid", 00:14:43.611 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:43.611 "strip_size_kb": 64, 00:14:43.611 "state": "configuring", 00:14:43.611 "raid_level": "raid5f", 00:14:43.611 "superblock": false, 00:14:43.611 "num_base_bdevs": 3, 00:14:43.611 "num_base_bdevs_discovered": 2, 00:14:43.611 "num_base_bdevs_operational": 3, 00:14:43.611 "base_bdevs_list": [ 00:14:43.611 { 00:14:43.611 "name": "BaseBdev1", 00:14:43.611 "uuid": "690cf7ac-b179-4ac5-8e53-018b22f92671", 00:14:43.611 "is_configured": true, 00:14:43.611 "data_offset": 0, 00:14:43.611 "data_size": 65536 00:14:43.611 }, 00:14:43.611 { 00:14:43.611 "name": null, 00:14:43.611 "uuid": "367d63c0-9aa4-46b1-a02b-a2171810fa26", 00:14:43.611 "is_configured": false, 00:14:43.611 "data_offset": 0, 00:14:43.611 "data_size": 65536 00:14:43.611 }, 00:14:43.611 { 00:14:43.611 "name": "BaseBdev3", 00:14:43.611 "uuid": "600f47a6-872d-4e13-89e6-a4a9e2b33425", 00:14:43.611 "is_configured": true, 00:14:43.611 "data_offset": 0, 00:14:43.611 "data_size": 65536 00:14:43.611 } 00:14:43.611 ] 00:14:43.611 }' 00:14:43.611 15:20:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:43.611 15:20:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:44.181 15:20:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:44.181 15:20:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:14:44.181 15:20:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:44.181 15:20:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:44.181 15:20:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:44.181 15:20:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:14:44.181 15:20:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:14:44.181 15:20:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:44.181 15:20:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:44.181 [2024-11-19 15:20:34.267280] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:44.181 15:20:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:44.181 15:20:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:44.181 15:20:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:44.181 15:20:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:44.181 15:20:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:44.181 15:20:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:44.181 15:20:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:44.181 15:20:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:44.181 15:20:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:44.181 15:20:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:44.181 15:20:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:44.181 15:20:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:44.181 15:20:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:44.181 15:20:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:44.181 15:20:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:44.181 15:20:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:44.181 15:20:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:44.181 "name": "Existed_Raid", 00:14:44.181 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:44.181 "strip_size_kb": 64, 00:14:44.181 "state": "configuring", 00:14:44.181 "raid_level": "raid5f", 00:14:44.181 "superblock": false, 00:14:44.181 "num_base_bdevs": 3, 00:14:44.181 "num_base_bdevs_discovered": 1, 00:14:44.181 "num_base_bdevs_operational": 3, 00:14:44.181 "base_bdevs_list": [ 00:14:44.181 { 00:14:44.181 "name": null, 00:14:44.182 "uuid": "690cf7ac-b179-4ac5-8e53-018b22f92671", 00:14:44.182 "is_configured": false, 00:14:44.182 "data_offset": 0, 00:14:44.182 "data_size": 65536 00:14:44.182 }, 00:14:44.182 { 00:14:44.182 "name": null, 00:14:44.182 "uuid": "367d63c0-9aa4-46b1-a02b-a2171810fa26", 00:14:44.182 "is_configured": false, 00:14:44.182 "data_offset": 0, 00:14:44.182 "data_size": 65536 00:14:44.182 }, 00:14:44.182 { 00:14:44.182 "name": "BaseBdev3", 00:14:44.182 "uuid": "600f47a6-872d-4e13-89e6-a4a9e2b33425", 00:14:44.182 "is_configured": true, 00:14:44.182 "data_offset": 0, 00:14:44.182 "data_size": 65536 00:14:44.182 } 00:14:44.182 ] 00:14:44.182 }' 00:14:44.182 15:20:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:44.182 15:20:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:44.442 15:20:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:44.442 15:20:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:14:44.442 15:20:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:44.442 15:20:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:44.442 15:20:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:44.442 15:20:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:14:44.442 15:20:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:14:44.442 15:20:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:44.442 15:20:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:44.442 [2024-11-19 15:20:34.713163] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:44.442 15:20:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:44.442 15:20:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:44.442 15:20:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:44.442 15:20:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:44.442 15:20:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:44.442 15:20:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:44.442 15:20:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:44.442 15:20:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:44.442 15:20:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:44.442 15:20:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:44.442 15:20:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:44.442 15:20:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:44.442 15:20:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:44.442 15:20:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:44.442 15:20:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:44.442 15:20:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:44.442 15:20:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:44.442 "name": "Existed_Raid", 00:14:44.442 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:44.442 "strip_size_kb": 64, 00:14:44.442 "state": "configuring", 00:14:44.442 "raid_level": "raid5f", 00:14:44.442 "superblock": false, 00:14:44.442 "num_base_bdevs": 3, 00:14:44.442 "num_base_bdevs_discovered": 2, 00:14:44.442 "num_base_bdevs_operational": 3, 00:14:44.442 "base_bdevs_list": [ 00:14:44.442 { 00:14:44.442 "name": null, 00:14:44.442 "uuid": "690cf7ac-b179-4ac5-8e53-018b22f92671", 00:14:44.442 "is_configured": false, 00:14:44.442 "data_offset": 0, 00:14:44.442 "data_size": 65536 00:14:44.442 }, 00:14:44.442 { 00:14:44.442 "name": "BaseBdev2", 00:14:44.442 "uuid": "367d63c0-9aa4-46b1-a02b-a2171810fa26", 00:14:44.442 "is_configured": true, 00:14:44.442 "data_offset": 0, 00:14:44.442 "data_size": 65536 00:14:44.442 }, 00:14:44.442 { 00:14:44.442 "name": "BaseBdev3", 00:14:44.442 "uuid": "600f47a6-872d-4e13-89e6-a4a9e2b33425", 00:14:44.442 "is_configured": true, 00:14:44.442 "data_offset": 0, 00:14:44.442 "data_size": 65536 00:14:44.442 } 00:14:44.442 ] 00:14:44.442 }' 00:14:44.442 15:20:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:44.442 15:20:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:45.013 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:14:45.013 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:45.013 15:20:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:45.013 15:20:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:45.013 15:20:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:45.013 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:14:45.013 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:45.013 15:20:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:45.013 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:14:45.013 15:20:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:45.013 15:20:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:45.013 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 690cf7ac-b179-4ac5-8e53-018b22f92671 00:14:45.013 15:20:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:45.013 15:20:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:45.013 [2024-11-19 15:20:35.219313] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:14:45.013 [2024-11-19 15:20:35.219428] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:14:45.013 [2024-11-19 15:20:35.219456] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:14:45.013 [2024-11-19 15:20:35.219715] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:14:45.013 [2024-11-19 15:20:35.220153] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:14:45.013 [2024-11-19 15:20:35.220202] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:14:45.013 [2024-11-19 15:20:35.220418] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:45.013 NewBaseBdev 00:14:45.013 15:20:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:45.013 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:14:45.013 15:20:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:14:45.013 15:20:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:45.013 15:20:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:14:45.013 15:20:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:45.013 15:20:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:45.013 15:20:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:45.013 15:20:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:45.013 15:20:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:45.013 15:20:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:45.013 15:20:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:14:45.013 15:20:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:45.013 15:20:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:45.013 [ 00:14:45.013 { 00:14:45.013 "name": "NewBaseBdev", 00:14:45.013 "aliases": [ 00:14:45.013 "690cf7ac-b179-4ac5-8e53-018b22f92671" 00:14:45.013 ], 00:14:45.013 "product_name": "Malloc disk", 00:14:45.013 "block_size": 512, 00:14:45.013 "num_blocks": 65536, 00:14:45.013 "uuid": "690cf7ac-b179-4ac5-8e53-018b22f92671", 00:14:45.013 "assigned_rate_limits": { 00:14:45.013 "rw_ios_per_sec": 0, 00:14:45.013 "rw_mbytes_per_sec": 0, 00:14:45.013 "r_mbytes_per_sec": 0, 00:14:45.013 "w_mbytes_per_sec": 0 00:14:45.013 }, 00:14:45.013 "claimed": true, 00:14:45.013 "claim_type": "exclusive_write", 00:14:45.013 "zoned": false, 00:14:45.013 "supported_io_types": { 00:14:45.013 "read": true, 00:14:45.013 "write": true, 00:14:45.013 "unmap": true, 00:14:45.013 "flush": true, 00:14:45.013 "reset": true, 00:14:45.013 "nvme_admin": false, 00:14:45.013 "nvme_io": false, 00:14:45.013 "nvme_io_md": false, 00:14:45.013 "write_zeroes": true, 00:14:45.013 "zcopy": true, 00:14:45.013 "get_zone_info": false, 00:14:45.013 "zone_management": false, 00:14:45.013 "zone_append": false, 00:14:45.013 "compare": false, 00:14:45.013 "compare_and_write": false, 00:14:45.013 "abort": true, 00:14:45.013 "seek_hole": false, 00:14:45.013 "seek_data": false, 00:14:45.013 "copy": true, 00:14:45.013 "nvme_iov_md": false 00:14:45.013 }, 00:14:45.013 "memory_domains": [ 00:14:45.013 { 00:14:45.013 "dma_device_id": "system", 00:14:45.013 "dma_device_type": 1 00:14:45.013 }, 00:14:45.013 { 00:14:45.013 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:45.014 "dma_device_type": 2 00:14:45.014 } 00:14:45.014 ], 00:14:45.014 "driver_specific": {} 00:14:45.014 } 00:14:45.014 ] 00:14:45.014 15:20:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:45.014 15:20:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:14:45.014 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:14:45.014 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:45.014 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:45.014 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:45.014 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:45.014 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:45.014 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:45.014 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:45.014 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:45.014 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:45.014 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:45.014 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:45.014 15:20:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:45.014 15:20:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:45.014 15:20:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:45.014 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:45.014 "name": "Existed_Raid", 00:14:45.014 "uuid": "99c9acb3-b341-4871-a038-f5630440a457", 00:14:45.014 "strip_size_kb": 64, 00:14:45.014 "state": "online", 00:14:45.014 "raid_level": "raid5f", 00:14:45.014 "superblock": false, 00:14:45.014 "num_base_bdevs": 3, 00:14:45.014 "num_base_bdevs_discovered": 3, 00:14:45.014 "num_base_bdevs_operational": 3, 00:14:45.014 "base_bdevs_list": [ 00:14:45.014 { 00:14:45.014 "name": "NewBaseBdev", 00:14:45.014 "uuid": "690cf7ac-b179-4ac5-8e53-018b22f92671", 00:14:45.014 "is_configured": true, 00:14:45.014 "data_offset": 0, 00:14:45.014 "data_size": 65536 00:14:45.014 }, 00:14:45.014 { 00:14:45.014 "name": "BaseBdev2", 00:14:45.014 "uuid": "367d63c0-9aa4-46b1-a02b-a2171810fa26", 00:14:45.014 "is_configured": true, 00:14:45.014 "data_offset": 0, 00:14:45.014 "data_size": 65536 00:14:45.014 }, 00:14:45.014 { 00:14:45.014 "name": "BaseBdev3", 00:14:45.014 "uuid": "600f47a6-872d-4e13-89e6-a4a9e2b33425", 00:14:45.014 "is_configured": true, 00:14:45.014 "data_offset": 0, 00:14:45.014 "data_size": 65536 00:14:45.014 } 00:14:45.014 ] 00:14:45.014 }' 00:14:45.014 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:45.014 15:20:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:45.584 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:14:45.584 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:14:45.584 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:45.584 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:45.584 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:14:45.584 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:45.584 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:14:45.584 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:45.584 15:20:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:45.584 15:20:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:45.584 [2024-11-19 15:20:35.730626] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:45.584 15:20:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:45.584 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:45.584 "name": "Existed_Raid", 00:14:45.584 "aliases": [ 00:14:45.584 "99c9acb3-b341-4871-a038-f5630440a457" 00:14:45.584 ], 00:14:45.584 "product_name": "Raid Volume", 00:14:45.584 "block_size": 512, 00:14:45.584 "num_blocks": 131072, 00:14:45.584 "uuid": "99c9acb3-b341-4871-a038-f5630440a457", 00:14:45.584 "assigned_rate_limits": { 00:14:45.584 "rw_ios_per_sec": 0, 00:14:45.584 "rw_mbytes_per_sec": 0, 00:14:45.584 "r_mbytes_per_sec": 0, 00:14:45.584 "w_mbytes_per_sec": 0 00:14:45.584 }, 00:14:45.584 "claimed": false, 00:14:45.584 "zoned": false, 00:14:45.584 "supported_io_types": { 00:14:45.584 "read": true, 00:14:45.584 "write": true, 00:14:45.584 "unmap": false, 00:14:45.584 "flush": false, 00:14:45.584 "reset": true, 00:14:45.584 "nvme_admin": false, 00:14:45.584 "nvme_io": false, 00:14:45.584 "nvme_io_md": false, 00:14:45.584 "write_zeroes": true, 00:14:45.584 "zcopy": false, 00:14:45.584 "get_zone_info": false, 00:14:45.584 "zone_management": false, 00:14:45.584 "zone_append": false, 00:14:45.584 "compare": false, 00:14:45.584 "compare_and_write": false, 00:14:45.584 "abort": false, 00:14:45.584 "seek_hole": false, 00:14:45.584 "seek_data": false, 00:14:45.584 "copy": false, 00:14:45.584 "nvme_iov_md": false 00:14:45.584 }, 00:14:45.584 "driver_specific": { 00:14:45.584 "raid": { 00:14:45.584 "uuid": "99c9acb3-b341-4871-a038-f5630440a457", 00:14:45.584 "strip_size_kb": 64, 00:14:45.584 "state": "online", 00:14:45.584 "raid_level": "raid5f", 00:14:45.584 "superblock": false, 00:14:45.584 "num_base_bdevs": 3, 00:14:45.584 "num_base_bdevs_discovered": 3, 00:14:45.584 "num_base_bdevs_operational": 3, 00:14:45.584 "base_bdevs_list": [ 00:14:45.584 { 00:14:45.584 "name": "NewBaseBdev", 00:14:45.584 "uuid": "690cf7ac-b179-4ac5-8e53-018b22f92671", 00:14:45.584 "is_configured": true, 00:14:45.584 "data_offset": 0, 00:14:45.584 "data_size": 65536 00:14:45.584 }, 00:14:45.584 { 00:14:45.584 "name": "BaseBdev2", 00:14:45.584 "uuid": "367d63c0-9aa4-46b1-a02b-a2171810fa26", 00:14:45.584 "is_configured": true, 00:14:45.584 "data_offset": 0, 00:14:45.584 "data_size": 65536 00:14:45.584 }, 00:14:45.584 { 00:14:45.584 "name": "BaseBdev3", 00:14:45.584 "uuid": "600f47a6-872d-4e13-89e6-a4a9e2b33425", 00:14:45.584 "is_configured": true, 00:14:45.584 "data_offset": 0, 00:14:45.584 "data_size": 65536 00:14:45.584 } 00:14:45.584 ] 00:14:45.584 } 00:14:45.584 } 00:14:45.584 }' 00:14:45.584 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:45.584 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:14:45.584 BaseBdev2 00:14:45.584 BaseBdev3' 00:14:45.584 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:45.584 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:45.584 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:45.584 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:14:45.584 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:45.584 15:20:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:45.584 15:20:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:45.584 15:20:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:45.584 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:45.584 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:45.584 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:45.584 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:14:45.584 15:20:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:45.584 15:20:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:45.844 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:45.844 15:20:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:45.844 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:45.844 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:45.844 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:45.844 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:45.844 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:14:45.844 15:20:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:45.844 15:20:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:45.844 15:20:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:45.844 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:45.845 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:45.845 15:20:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:45.845 15:20:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:45.845 15:20:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:45.845 [2024-11-19 15:20:36.005993] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:45.845 [2024-11-19 15:20:36.006054] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:45.845 [2024-11-19 15:20:36.006163] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:45.845 [2024-11-19 15:20:36.006426] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:45.845 [2024-11-19 15:20:36.006478] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:14:45.845 15:20:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:45.845 15:20:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 90462 00:14:45.845 15:20:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 90462 ']' 00:14:45.845 15:20:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@958 -- # kill -0 90462 00:14:45.845 15:20:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@959 -- # uname 00:14:45.845 15:20:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:14:45.845 15:20:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 90462 00:14:45.845 15:20:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:14:45.845 15:20:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:14:45.845 15:20:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 90462' 00:14:45.845 killing process with pid 90462 00:14:45.845 15:20:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@973 -- # kill 90462 00:14:45.845 [2024-11-19 15:20:36.042413] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:45.845 15:20:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@978 -- # wait 90462 00:14:45.845 [2024-11-19 15:20:36.074005] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:46.105 15:20:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:14:46.105 00:14:46.105 real 0m8.904s 00:14:46.105 user 0m15.214s 00:14:46.105 sys 0m1.923s 00:14:46.105 15:20:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:14:46.105 ************************************ 00:14:46.105 END TEST raid5f_state_function_test 00:14:46.105 ************************************ 00:14:46.105 15:20:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:46.105 15:20:36 bdev_raid -- bdev/bdev_raid.sh@987 -- # run_test raid5f_state_function_test_sb raid_state_function_test raid5f 3 true 00:14:46.105 15:20:36 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:14:46.105 15:20:36 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:14:46.105 15:20:36 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:46.105 ************************************ 00:14:46.105 START TEST raid5f_state_function_test_sb 00:14:46.105 ************************************ 00:14:46.105 15:20:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid5f 3 true 00:14:46.105 15:20:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:14:46.105 15:20:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:14:46.105 15:20:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:14:46.105 15:20:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:14:46.105 15:20:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:14:46.105 15:20:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:46.105 15:20:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:14:46.105 15:20:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:46.105 15:20:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:46.105 15:20:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:14:46.105 15:20:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:46.105 15:20:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:46.105 15:20:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:14:46.105 15:20:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:46.105 15:20:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:46.105 15:20:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:14:46.105 15:20:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:14:46.105 15:20:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:14:46.105 15:20:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:14:46.105 15:20:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:14:46.105 15:20:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:14:46.105 15:20:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:14:46.105 15:20:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:14:46.105 15:20:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:14:46.105 15:20:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:14:46.105 15:20:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:14:46.105 Process raid pid: 91067 00:14:46.105 15:20:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=91067 00:14:46.105 15:20:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:14:46.105 15:20:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 91067' 00:14:46.105 15:20:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 91067 00:14:46.105 15:20:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 91067 ']' 00:14:46.105 15:20:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:46.105 15:20:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:14:46.105 15:20:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:46.105 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:46.105 15:20:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:14:46.105 15:20:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:46.366 [2024-11-19 15:20:36.478331] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:14:46.366 [2024-11-19 15:20:36.478470] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:14:46.366 [2024-11-19 15:20:36.615162] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:46.366 [2024-11-19 15:20:36.639679] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:14:46.366 [2024-11-19 15:20:36.682688] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:46.366 [2024-11-19 15:20:36.682804] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:47.306 15:20:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:14:47.306 15:20:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:14:47.306 15:20:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:14:47.306 15:20:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:47.306 15:20:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:47.306 [2024-11-19 15:20:37.288667] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:47.306 [2024-11-19 15:20:37.288783] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:47.306 [2024-11-19 15:20:37.288816] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:47.306 [2024-11-19 15:20:37.288840] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:47.306 [2024-11-19 15:20:37.288858] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:47.306 [2024-11-19 15:20:37.288880] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:47.306 15:20:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:47.306 15:20:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:47.306 15:20:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:47.306 15:20:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:47.306 15:20:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:47.306 15:20:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:47.306 15:20:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:47.306 15:20:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:47.306 15:20:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:47.306 15:20:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:47.306 15:20:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:47.306 15:20:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:47.306 15:20:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:47.306 15:20:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:47.306 15:20:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:47.306 15:20:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:47.306 15:20:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:47.306 "name": "Existed_Raid", 00:14:47.306 "uuid": "6dafd658-744a-4885-bbe2-62247adaf193", 00:14:47.306 "strip_size_kb": 64, 00:14:47.306 "state": "configuring", 00:14:47.306 "raid_level": "raid5f", 00:14:47.306 "superblock": true, 00:14:47.306 "num_base_bdevs": 3, 00:14:47.306 "num_base_bdevs_discovered": 0, 00:14:47.306 "num_base_bdevs_operational": 3, 00:14:47.306 "base_bdevs_list": [ 00:14:47.306 { 00:14:47.306 "name": "BaseBdev1", 00:14:47.306 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:47.306 "is_configured": false, 00:14:47.306 "data_offset": 0, 00:14:47.306 "data_size": 0 00:14:47.306 }, 00:14:47.306 { 00:14:47.306 "name": "BaseBdev2", 00:14:47.306 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:47.306 "is_configured": false, 00:14:47.306 "data_offset": 0, 00:14:47.306 "data_size": 0 00:14:47.306 }, 00:14:47.306 { 00:14:47.306 "name": "BaseBdev3", 00:14:47.306 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:47.306 "is_configured": false, 00:14:47.306 "data_offset": 0, 00:14:47.306 "data_size": 0 00:14:47.306 } 00:14:47.306 ] 00:14:47.306 }' 00:14:47.306 15:20:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:47.306 15:20:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:47.567 15:20:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:47.567 15:20:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:47.567 15:20:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:47.567 [2024-11-19 15:20:37.755863] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:47.567 [2024-11-19 15:20:37.755946] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:14:47.567 15:20:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:47.567 15:20:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:14:47.567 15:20:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:47.567 15:20:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:47.567 [2024-11-19 15:20:37.763883] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:47.567 [2024-11-19 15:20:37.763923] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:47.567 [2024-11-19 15:20:37.763932] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:47.567 [2024-11-19 15:20:37.763941] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:47.567 [2024-11-19 15:20:37.763947] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:47.567 [2024-11-19 15:20:37.763956] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:47.567 15:20:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:47.567 15:20:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:14:47.567 15:20:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:47.567 15:20:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:47.567 [2024-11-19 15:20:37.780941] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:47.567 BaseBdev1 00:14:47.567 15:20:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:47.567 15:20:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:14:47.567 15:20:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:14:47.567 15:20:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:47.567 15:20:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:14:47.567 15:20:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:47.567 15:20:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:47.567 15:20:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:47.567 15:20:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:47.567 15:20:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:47.567 15:20:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:47.567 15:20:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:14:47.567 15:20:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:47.567 15:20:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:47.567 [ 00:14:47.567 { 00:14:47.567 "name": "BaseBdev1", 00:14:47.567 "aliases": [ 00:14:47.567 "630d6d6a-fa59-45d5-bfae-9be1bfb92353" 00:14:47.567 ], 00:14:47.567 "product_name": "Malloc disk", 00:14:47.567 "block_size": 512, 00:14:47.567 "num_blocks": 65536, 00:14:47.567 "uuid": "630d6d6a-fa59-45d5-bfae-9be1bfb92353", 00:14:47.567 "assigned_rate_limits": { 00:14:47.567 "rw_ios_per_sec": 0, 00:14:47.567 "rw_mbytes_per_sec": 0, 00:14:47.567 "r_mbytes_per_sec": 0, 00:14:47.567 "w_mbytes_per_sec": 0 00:14:47.567 }, 00:14:47.567 "claimed": true, 00:14:47.567 "claim_type": "exclusive_write", 00:14:47.567 "zoned": false, 00:14:47.567 "supported_io_types": { 00:14:47.567 "read": true, 00:14:47.567 "write": true, 00:14:47.567 "unmap": true, 00:14:47.567 "flush": true, 00:14:47.567 "reset": true, 00:14:47.567 "nvme_admin": false, 00:14:47.567 "nvme_io": false, 00:14:47.567 "nvme_io_md": false, 00:14:47.567 "write_zeroes": true, 00:14:47.567 "zcopy": true, 00:14:47.567 "get_zone_info": false, 00:14:47.567 "zone_management": false, 00:14:47.567 "zone_append": false, 00:14:47.567 "compare": false, 00:14:47.567 "compare_and_write": false, 00:14:47.567 "abort": true, 00:14:47.567 "seek_hole": false, 00:14:47.567 "seek_data": false, 00:14:47.567 "copy": true, 00:14:47.567 "nvme_iov_md": false 00:14:47.567 }, 00:14:47.567 "memory_domains": [ 00:14:47.567 { 00:14:47.567 "dma_device_id": "system", 00:14:47.567 "dma_device_type": 1 00:14:47.567 }, 00:14:47.567 { 00:14:47.567 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:47.567 "dma_device_type": 2 00:14:47.567 } 00:14:47.567 ], 00:14:47.567 "driver_specific": {} 00:14:47.567 } 00:14:47.567 ] 00:14:47.567 15:20:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:47.567 15:20:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:14:47.567 15:20:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:47.568 15:20:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:47.568 15:20:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:47.568 15:20:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:47.568 15:20:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:47.568 15:20:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:47.568 15:20:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:47.568 15:20:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:47.568 15:20:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:47.568 15:20:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:47.568 15:20:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:47.568 15:20:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:47.568 15:20:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:47.568 15:20:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:47.568 15:20:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:47.568 15:20:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:47.568 "name": "Existed_Raid", 00:14:47.568 "uuid": "63c35e37-5dc0-414c-8ac3-3e803cea0c7a", 00:14:47.568 "strip_size_kb": 64, 00:14:47.568 "state": "configuring", 00:14:47.568 "raid_level": "raid5f", 00:14:47.568 "superblock": true, 00:14:47.568 "num_base_bdevs": 3, 00:14:47.568 "num_base_bdevs_discovered": 1, 00:14:47.568 "num_base_bdevs_operational": 3, 00:14:47.568 "base_bdevs_list": [ 00:14:47.568 { 00:14:47.568 "name": "BaseBdev1", 00:14:47.568 "uuid": "630d6d6a-fa59-45d5-bfae-9be1bfb92353", 00:14:47.568 "is_configured": true, 00:14:47.568 "data_offset": 2048, 00:14:47.568 "data_size": 63488 00:14:47.568 }, 00:14:47.568 { 00:14:47.568 "name": "BaseBdev2", 00:14:47.568 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:47.568 "is_configured": false, 00:14:47.568 "data_offset": 0, 00:14:47.568 "data_size": 0 00:14:47.568 }, 00:14:47.568 { 00:14:47.568 "name": "BaseBdev3", 00:14:47.568 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:47.568 "is_configured": false, 00:14:47.568 "data_offset": 0, 00:14:47.568 "data_size": 0 00:14:47.568 } 00:14:47.568 ] 00:14:47.568 }' 00:14:47.568 15:20:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:47.568 15:20:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:48.138 15:20:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:48.138 15:20:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:48.138 15:20:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:48.138 [2024-11-19 15:20:38.244145] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:48.139 [2024-11-19 15:20:38.244240] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:14:48.139 15:20:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:48.139 15:20:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:14:48.139 15:20:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:48.139 15:20:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:48.139 [2024-11-19 15:20:38.256175] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:48.139 [2024-11-19 15:20:38.258024] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:48.139 [2024-11-19 15:20:38.258093] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:48.139 [2024-11-19 15:20:38.258136] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:48.139 [2024-11-19 15:20:38.258160] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:48.139 15:20:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:48.139 15:20:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:14:48.139 15:20:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:48.139 15:20:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:48.139 15:20:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:48.139 15:20:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:48.139 15:20:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:48.139 15:20:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:48.139 15:20:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:48.139 15:20:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:48.139 15:20:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:48.139 15:20:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:48.139 15:20:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:48.139 15:20:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:48.139 15:20:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:48.139 15:20:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:48.139 15:20:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:48.139 15:20:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:48.139 15:20:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:48.139 "name": "Existed_Raid", 00:14:48.139 "uuid": "dccb2ffd-9fe9-4629-b9a3-122890389dae", 00:14:48.139 "strip_size_kb": 64, 00:14:48.139 "state": "configuring", 00:14:48.139 "raid_level": "raid5f", 00:14:48.139 "superblock": true, 00:14:48.139 "num_base_bdevs": 3, 00:14:48.139 "num_base_bdevs_discovered": 1, 00:14:48.139 "num_base_bdevs_operational": 3, 00:14:48.139 "base_bdevs_list": [ 00:14:48.139 { 00:14:48.139 "name": "BaseBdev1", 00:14:48.139 "uuid": "630d6d6a-fa59-45d5-bfae-9be1bfb92353", 00:14:48.139 "is_configured": true, 00:14:48.139 "data_offset": 2048, 00:14:48.139 "data_size": 63488 00:14:48.139 }, 00:14:48.139 { 00:14:48.139 "name": "BaseBdev2", 00:14:48.139 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:48.139 "is_configured": false, 00:14:48.139 "data_offset": 0, 00:14:48.139 "data_size": 0 00:14:48.139 }, 00:14:48.139 { 00:14:48.139 "name": "BaseBdev3", 00:14:48.139 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:48.139 "is_configured": false, 00:14:48.139 "data_offset": 0, 00:14:48.139 "data_size": 0 00:14:48.139 } 00:14:48.139 ] 00:14:48.139 }' 00:14:48.139 15:20:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:48.139 15:20:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:48.399 15:20:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:14:48.399 15:20:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:48.399 15:20:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:48.399 [2024-11-19 15:20:38.706474] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:48.399 BaseBdev2 00:14:48.399 15:20:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:48.399 15:20:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:14:48.399 15:20:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:14:48.399 15:20:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:48.399 15:20:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:14:48.399 15:20:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:48.400 15:20:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:48.400 15:20:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:48.400 15:20:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:48.400 15:20:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:48.400 15:20:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:48.400 15:20:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:14:48.400 15:20:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:48.400 15:20:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:48.400 [ 00:14:48.400 { 00:14:48.400 "name": "BaseBdev2", 00:14:48.400 "aliases": [ 00:14:48.400 "e0aee967-7675-41cb-97ab-629ed897346e" 00:14:48.400 ], 00:14:48.400 "product_name": "Malloc disk", 00:14:48.400 "block_size": 512, 00:14:48.400 "num_blocks": 65536, 00:14:48.400 "uuid": "e0aee967-7675-41cb-97ab-629ed897346e", 00:14:48.400 "assigned_rate_limits": { 00:14:48.400 "rw_ios_per_sec": 0, 00:14:48.400 "rw_mbytes_per_sec": 0, 00:14:48.400 "r_mbytes_per_sec": 0, 00:14:48.400 "w_mbytes_per_sec": 0 00:14:48.400 }, 00:14:48.400 "claimed": true, 00:14:48.400 "claim_type": "exclusive_write", 00:14:48.400 "zoned": false, 00:14:48.400 "supported_io_types": { 00:14:48.400 "read": true, 00:14:48.400 "write": true, 00:14:48.400 "unmap": true, 00:14:48.400 "flush": true, 00:14:48.400 "reset": true, 00:14:48.400 "nvme_admin": false, 00:14:48.400 "nvme_io": false, 00:14:48.400 "nvme_io_md": false, 00:14:48.400 "write_zeroes": true, 00:14:48.400 "zcopy": true, 00:14:48.660 "get_zone_info": false, 00:14:48.660 "zone_management": false, 00:14:48.660 "zone_append": false, 00:14:48.660 "compare": false, 00:14:48.660 "compare_and_write": false, 00:14:48.660 "abort": true, 00:14:48.660 "seek_hole": false, 00:14:48.660 "seek_data": false, 00:14:48.660 "copy": true, 00:14:48.660 "nvme_iov_md": false 00:14:48.660 }, 00:14:48.660 "memory_domains": [ 00:14:48.660 { 00:14:48.660 "dma_device_id": "system", 00:14:48.660 "dma_device_type": 1 00:14:48.660 }, 00:14:48.660 { 00:14:48.660 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:48.660 "dma_device_type": 2 00:14:48.660 } 00:14:48.660 ], 00:14:48.660 "driver_specific": {} 00:14:48.660 } 00:14:48.660 ] 00:14:48.660 15:20:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:48.660 15:20:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:14:48.660 15:20:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:48.660 15:20:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:48.660 15:20:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:48.660 15:20:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:48.660 15:20:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:48.660 15:20:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:48.660 15:20:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:48.660 15:20:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:48.660 15:20:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:48.660 15:20:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:48.660 15:20:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:48.660 15:20:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:48.660 15:20:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:48.660 15:20:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:48.660 15:20:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:48.660 15:20:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:48.660 15:20:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:48.660 15:20:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:48.660 "name": "Existed_Raid", 00:14:48.660 "uuid": "dccb2ffd-9fe9-4629-b9a3-122890389dae", 00:14:48.660 "strip_size_kb": 64, 00:14:48.660 "state": "configuring", 00:14:48.660 "raid_level": "raid5f", 00:14:48.660 "superblock": true, 00:14:48.660 "num_base_bdevs": 3, 00:14:48.660 "num_base_bdevs_discovered": 2, 00:14:48.660 "num_base_bdevs_operational": 3, 00:14:48.660 "base_bdevs_list": [ 00:14:48.660 { 00:14:48.660 "name": "BaseBdev1", 00:14:48.661 "uuid": "630d6d6a-fa59-45d5-bfae-9be1bfb92353", 00:14:48.661 "is_configured": true, 00:14:48.661 "data_offset": 2048, 00:14:48.661 "data_size": 63488 00:14:48.661 }, 00:14:48.661 { 00:14:48.661 "name": "BaseBdev2", 00:14:48.661 "uuid": "e0aee967-7675-41cb-97ab-629ed897346e", 00:14:48.661 "is_configured": true, 00:14:48.661 "data_offset": 2048, 00:14:48.661 "data_size": 63488 00:14:48.661 }, 00:14:48.661 { 00:14:48.661 "name": "BaseBdev3", 00:14:48.661 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:48.661 "is_configured": false, 00:14:48.661 "data_offset": 0, 00:14:48.661 "data_size": 0 00:14:48.661 } 00:14:48.661 ] 00:14:48.661 }' 00:14:48.661 15:20:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:48.661 15:20:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:48.921 15:20:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:14:48.921 15:20:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:48.921 15:20:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:48.921 [2024-11-19 15:20:39.218568] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:48.921 [2024-11-19 15:20:39.218784] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:14:48.921 [2024-11-19 15:20:39.218804] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:14:48.921 [2024-11-19 15:20:39.219156] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:14:48.921 BaseBdev3 00:14:48.921 [2024-11-19 15:20:39.219658] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:14:48.921 [2024-11-19 15:20:39.219684] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:14:48.921 [2024-11-19 15:20:39.219844] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:48.921 15:20:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:48.921 15:20:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:14:48.921 15:20:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:14:48.921 15:20:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:48.921 15:20:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:14:48.921 15:20:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:48.921 15:20:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:48.921 15:20:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:48.921 15:20:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:48.921 15:20:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:48.921 15:20:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:48.921 15:20:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:14:48.921 15:20:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:48.921 15:20:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:48.921 [ 00:14:48.921 { 00:14:48.921 "name": "BaseBdev3", 00:14:48.921 "aliases": [ 00:14:48.921 "68f29edc-a798-49c0-a74a-e194e522c2e8" 00:14:48.921 ], 00:14:48.921 "product_name": "Malloc disk", 00:14:48.921 "block_size": 512, 00:14:48.921 "num_blocks": 65536, 00:14:48.921 "uuid": "68f29edc-a798-49c0-a74a-e194e522c2e8", 00:14:48.921 "assigned_rate_limits": { 00:14:48.921 "rw_ios_per_sec": 0, 00:14:48.921 "rw_mbytes_per_sec": 0, 00:14:48.921 "r_mbytes_per_sec": 0, 00:14:48.921 "w_mbytes_per_sec": 0 00:14:48.921 }, 00:14:48.921 "claimed": true, 00:14:48.921 "claim_type": "exclusive_write", 00:14:48.921 "zoned": false, 00:14:48.921 "supported_io_types": { 00:14:48.921 "read": true, 00:14:48.921 "write": true, 00:14:48.921 "unmap": true, 00:14:48.921 "flush": true, 00:14:48.921 "reset": true, 00:14:48.921 "nvme_admin": false, 00:14:48.921 "nvme_io": false, 00:14:48.921 "nvme_io_md": false, 00:14:48.921 "write_zeroes": true, 00:14:48.921 "zcopy": true, 00:14:48.921 "get_zone_info": false, 00:14:48.921 "zone_management": false, 00:14:48.921 "zone_append": false, 00:14:48.921 "compare": false, 00:14:48.921 "compare_and_write": false, 00:14:48.921 "abort": true, 00:14:48.921 "seek_hole": false, 00:14:48.921 "seek_data": false, 00:14:48.921 "copy": true, 00:14:48.921 "nvme_iov_md": false 00:14:48.921 }, 00:14:48.921 "memory_domains": [ 00:14:48.921 { 00:14:48.921 "dma_device_id": "system", 00:14:48.921 "dma_device_type": 1 00:14:48.921 }, 00:14:48.921 { 00:14:48.921 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:48.921 "dma_device_type": 2 00:14:48.921 } 00:14:48.921 ], 00:14:48.921 "driver_specific": {} 00:14:48.921 } 00:14:48.921 ] 00:14:48.921 15:20:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:48.921 15:20:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:14:48.921 15:20:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:48.921 15:20:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:48.921 15:20:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:14:48.921 15:20:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:48.921 15:20:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:48.921 15:20:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:48.922 15:20:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:48.922 15:20:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:49.181 15:20:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:49.181 15:20:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:49.181 15:20:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:49.181 15:20:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:49.181 15:20:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:49.181 15:20:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:49.181 15:20:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.181 15:20:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:49.181 15:20:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.181 15:20:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:49.181 "name": "Existed_Raid", 00:14:49.181 "uuid": "dccb2ffd-9fe9-4629-b9a3-122890389dae", 00:14:49.181 "strip_size_kb": 64, 00:14:49.181 "state": "online", 00:14:49.181 "raid_level": "raid5f", 00:14:49.181 "superblock": true, 00:14:49.181 "num_base_bdevs": 3, 00:14:49.181 "num_base_bdevs_discovered": 3, 00:14:49.181 "num_base_bdevs_operational": 3, 00:14:49.181 "base_bdevs_list": [ 00:14:49.181 { 00:14:49.181 "name": "BaseBdev1", 00:14:49.181 "uuid": "630d6d6a-fa59-45d5-bfae-9be1bfb92353", 00:14:49.181 "is_configured": true, 00:14:49.181 "data_offset": 2048, 00:14:49.181 "data_size": 63488 00:14:49.181 }, 00:14:49.181 { 00:14:49.181 "name": "BaseBdev2", 00:14:49.181 "uuid": "e0aee967-7675-41cb-97ab-629ed897346e", 00:14:49.181 "is_configured": true, 00:14:49.181 "data_offset": 2048, 00:14:49.181 "data_size": 63488 00:14:49.181 }, 00:14:49.181 { 00:14:49.181 "name": "BaseBdev3", 00:14:49.181 "uuid": "68f29edc-a798-49c0-a74a-e194e522c2e8", 00:14:49.181 "is_configured": true, 00:14:49.181 "data_offset": 2048, 00:14:49.181 "data_size": 63488 00:14:49.181 } 00:14:49.181 ] 00:14:49.181 }' 00:14:49.181 15:20:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:49.181 15:20:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:49.442 15:20:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:14:49.442 15:20:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:14:49.442 15:20:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:49.442 15:20:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:49.442 15:20:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:14:49.442 15:20:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:49.442 15:20:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:14:49.442 15:20:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:49.442 15:20:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.442 15:20:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:49.442 [2024-11-19 15:20:39.729890] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:49.442 15:20:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.442 15:20:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:49.442 "name": "Existed_Raid", 00:14:49.442 "aliases": [ 00:14:49.442 "dccb2ffd-9fe9-4629-b9a3-122890389dae" 00:14:49.442 ], 00:14:49.442 "product_name": "Raid Volume", 00:14:49.442 "block_size": 512, 00:14:49.442 "num_blocks": 126976, 00:14:49.442 "uuid": "dccb2ffd-9fe9-4629-b9a3-122890389dae", 00:14:49.442 "assigned_rate_limits": { 00:14:49.442 "rw_ios_per_sec": 0, 00:14:49.442 "rw_mbytes_per_sec": 0, 00:14:49.442 "r_mbytes_per_sec": 0, 00:14:49.442 "w_mbytes_per_sec": 0 00:14:49.442 }, 00:14:49.442 "claimed": false, 00:14:49.442 "zoned": false, 00:14:49.442 "supported_io_types": { 00:14:49.442 "read": true, 00:14:49.442 "write": true, 00:14:49.442 "unmap": false, 00:14:49.442 "flush": false, 00:14:49.442 "reset": true, 00:14:49.442 "nvme_admin": false, 00:14:49.442 "nvme_io": false, 00:14:49.442 "nvme_io_md": false, 00:14:49.442 "write_zeroes": true, 00:14:49.442 "zcopy": false, 00:14:49.442 "get_zone_info": false, 00:14:49.442 "zone_management": false, 00:14:49.442 "zone_append": false, 00:14:49.442 "compare": false, 00:14:49.442 "compare_and_write": false, 00:14:49.442 "abort": false, 00:14:49.442 "seek_hole": false, 00:14:49.442 "seek_data": false, 00:14:49.442 "copy": false, 00:14:49.442 "nvme_iov_md": false 00:14:49.442 }, 00:14:49.442 "driver_specific": { 00:14:49.442 "raid": { 00:14:49.442 "uuid": "dccb2ffd-9fe9-4629-b9a3-122890389dae", 00:14:49.442 "strip_size_kb": 64, 00:14:49.442 "state": "online", 00:14:49.442 "raid_level": "raid5f", 00:14:49.442 "superblock": true, 00:14:49.442 "num_base_bdevs": 3, 00:14:49.442 "num_base_bdevs_discovered": 3, 00:14:49.442 "num_base_bdevs_operational": 3, 00:14:49.442 "base_bdevs_list": [ 00:14:49.442 { 00:14:49.442 "name": "BaseBdev1", 00:14:49.442 "uuid": "630d6d6a-fa59-45d5-bfae-9be1bfb92353", 00:14:49.442 "is_configured": true, 00:14:49.442 "data_offset": 2048, 00:14:49.442 "data_size": 63488 00:14:49.442 }, 00:14:49.442 { 00:14:49.442 "name": "BaseBdev2", 00:14:49.442 "uuid": "e0aee967-7675-41cb-97ab-629ed897346e", 00:14:49.442 "is_configured": true, 00:14:49.442 "data_offset": 2048, 00:14:49.442 "data_size": 63488 00:14:49.442 }, 00:14:49.442 { 00:14:49.442 "name": "BaseBdev3", 00:14:49.442 "uuid": "68f29edc-a798-49c0-a74a-e194e522c2e8", 00:14:49.442 "is_configured": true, 00:14:49.442 "data_offset": 2048, 00:14:49.442 "data_size": 63488 00:14:49.442 } 00:14:49.442 ] 00:14:49.442 } 00:14:49.442 } 00:14:49.442 }' 00:14:49.442 15:20:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:49.703 15:20:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:14:49.703 BaseBdev2 00:14:49.703 BaseBdev3' 00:14:49.703 15:20:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:49.703 15:20:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:49.703 15:20:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:49.703 15:20:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:14:49.703 15:20:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.703 15:20:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:49.703 15:20:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:49.703 15:20:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.703 15:20:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:49.703 15:20:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:49.703 15:20:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:49.703 15:20:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:49.703 15:20:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:14:49.703 15:20:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.703 15:20:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:49.703 15:20:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.703 15:20:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:49.703 15:20:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:49.703 15:20:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:49.703 15:20:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:14:49.703 15:20:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:49.703 15:20:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.703 15:20:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:49.703 15:20:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.703 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:49.703 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:49.703 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:14:49.703 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.703 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:49.703 [2024-11-19 15:20:40.009275] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:49.703 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.703 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:14:49.703 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:14:49.703 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:14:49.703 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:14:49.703 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:14:49.703 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 2 00:14:49.703 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:49.703 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:49.703 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:49.703 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:49.703 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:49.703 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:49.703 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:49.703 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:49.703 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:49.703 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:49.703 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:49.703 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.703 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:49.963 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.963 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:49.963 "name": "Existed_Raid", 00:14:49.963 "uuid": "dccb2ffd-9fe9-4629-b9a3-122890389dae", 00:14:49.963 "strip_size_kb": 64, 00:14:49.963 "state": "online", 00:14:49.963 "raid_level": "raid5f", 00:14:49.963 "superblock": true, 00:14:49.963 "num_base_bdevs": 3, 00:14:49.963 "num_base_bdevs_discovered": 2, 00:14:49.963 "num_base_bdevs_operational": 2, 00:14:49.963 "base_bdevs_list": [ 00:14:49.963 { 00:14:49.963 "name": null, 00:14:49.963 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:49.963 "is_configured": false, 00:14:49.963 "data_offset": 0, 00:14:49.963 "data_size": 63488 00:14:49.963 }, 00:14:49.963 { 00:14:49.963 "name": "BaseBdev2", 00:14:49.963 "uuid": "e0aee967-7675-41cb-97ab-629ed897346e", 00:14:49.963 "is_configured": true, 00:14:49.963 "data_offset": 2048, 00:14:49.963 "data_size": 63488 00:14:49.963 }, 00:14:49.963 { 00:14:49.963 "name": "BaseBdev3", 00:14:49.963 "uuid": "68f29edc-a798-49c0-a74a-e194e522c2e8", 00:14:49.963 "is_configured": true, 00:14:49.963 "data_offset": 2048, 00:14:49.963 "data_size": 63488 00:14:49.963 } 00:14:49.963 ] 00:14:49.963 }' 00:14:49.963 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:49.963 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:50.223 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:14:50.223 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:50.223 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:50.223 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:50.223 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:50.223 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:50.223 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:50.223 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:50.223 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:50.223 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:14:50.223 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:50.223 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:50.223 [2024-11-19 15:20:40.491928] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:14:50.223 [2024-11-19 15:20:40.492064] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:50.223 [2024-11-19 15:20:40.503182] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:50.223 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:50.223 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:50.223 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:50.223 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:50.223 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:50.223 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:50.223 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:50.223 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:50.223 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:50.223 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:50.223 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:14:50.223 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:50.223 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:50.484 [2024-11-19 15:20:40.563098] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:14:50.484 [2024-11-19 15:20:40.563139] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:14:50.484 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:50.484 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:50.484 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:50.484 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:50.484 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:14:50.484 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:50.484 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:50.484 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:50.484 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:14:50.484 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:14:50.484 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:14:50.484 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:14:50.484 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:50.484 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:14:50.484 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:50.484 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:50.484 BaseBdev2 00:14:50.484 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:50.484 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:14:50.484 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:14:50.484 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:50.484 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:14:50.484 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:50.484 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:50.484 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:50.484 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:50.484 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:50.484 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:50.484 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:14:50.484 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:50.484 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:50.484 [ 00:14:50.484 { 00:14:50.484 "name": "BaseBdev2", 00:14:50.484 "aliases": [ 00:14:50.484 "f24d9346-c89f-4f7e-a1f2-9743f7d7f66f" 00:14:50.484 ], 00:14:50.484 "product_name": "Malloc disk", 00:14:50.484 "block_size": 512, 00:14:50.484 "num_blocks": 65536, 00:14:50.484 "uuid": "f24d9346-c89f-4f7e-a1f2-9743f7d7f66f", 00:14:50.484 "assigned_rate_limits": { 00:14:50.484 "rw_ios_per_sec": 0, 00:14:50.484 "rw_mbytes_per_sec": 0, 00:14:50.484 "r_mbytes_per_sec": 0, 00:14:50.484 "w_mbytes_per_sec": 0 00:14:50.484 }, 00:14:50.484 "claimed": false, 00:14:50.484 "zoned": false, 00:14:50.484 "supported_io_types": { 00:14:50.484 "read": true, 00:14:50.484 "write": true, 00:14:50.484 "unmap": true, 00:14:50.484 "flush": true, 00:14:50.484 "reset": true, 00:14:50.484 "nvme_admin": false, 00:14:50.484 "nvme_io": false, 00:14:50.484 "nvme_io_md": false, 00:14:50.484 "write_zeroes": true, 00:14:50.484 "zcopy": true, 00:14:50.484 "get_zone_info": false, 00:14:50.484 "zone_management": false, 00:14:50.484 "zone_append": false, 00:14:50.484 "compare": false, 00:14:50.484 "compare_and_write": false, 00:14:50.484 "abort": true, 00:14:50.484 "seek_hole": false, 00:14:50.484 "seek_data": false, 00:14:50.484 "copy": true, 00:14:50.484 "nvme_iov_md": false 00:14:50.484 }, 00:14:50.484 "memory_domains": [ 00:14:50.484 { 00:14:50.484 "dma_device_id": "system", 00:14:50.484 "dma_device_type": 1 00:14:50.484 }, 00:14:50.484 { 00:14:50.484 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:50.484 "dma_device_type": 2 00:14:50.484 } 00:14:50.484 ], 00:14:50.484 "driver_specific": {} 00:14:50.484 } 00:14:50.484 ] 00:14:50.484 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:50.484 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:14:50.484 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:50.484 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:50.484 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:14:50.484 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:50.484 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:50.484 BaseBdev3 00:14:50.484 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:50.484 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:14:50.484 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:14:50.484 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:50.484 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:14:50.484 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:50.484 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:50.485 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:50.485 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:50.485 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:50.485 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:50.485 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:14:50.485 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:50.485 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:50.485 [ 00:14:50.485 { 00:14:50.485 "name": "BaseBdev3", 00:14:50.485 "aliases": [ 00:14:50.485 "6e089d12-83f5-4d78-ba3b-e672cda93bf0" 00:14:50.485 ], 00:14:50.485 "product_name": "Malloc disk", 00:14:50.485 "block_size": 512, 00:14:50.485 "num_blocks": 65536, 00:14:50.485 "uuid": "6e089d12-83f5-4d78-ba3b-e672cda93bf0", 00:14:50.485 "assigned_rate_limits": { 00:14:50.485 "rw_ios_per_sec": 0, 00:14:50.485 "rw_mbytes_per_sec": 0, 00:14:50.485 "r_mbytes_per_sec": 0, 00:14:50.485 "w_mbytes_per_sec": 0 00:14:50.485 }, 00:14:50.485 "claimed": false, 00:14:50.485 "zoned": false, 00:14:50.485 "supported_io_types": { 00:14:50.485 "read": true, 00:14:50.485 "write": true, 00:14:50.485 "unmap": true, 00:14:50.485 "flush": true, 00:14:50.485 "reset": true, 00:14:50.485 "nvme_admin": false, 00:14:50.485 "nvme_io": false, 00:14:50.485 "nvme_io_md": false, 00:14:50.485 "write_zeroes": true, 00:14:50.485 "zcopy": true, 00:14:50.485 "get_zone_info": false, 00:14:50.485 "zone_management": false, 00:14:50.485 "zone_append": false, 00:14:50.485 "compare": false, 00:14:50.485 "compare_and_write": false, 00:14:50.485 "abort": true, 00:14:50.485 "seek_hole": false, 00:14:50.485 "seek_data": false, 00:14:50.485 "copy": true, 00:14:50.485 "nvme_iov_md": false 00:14:50.485 }, 00:14:50.485 "memory_domains": [ 00:14:50.485 { 00:14:50.485 "dma_device_id": "system", 00:14:50.485 "dma_device_type": 1 00:14:50.485 }, 00:14:50.485 { 00:14:50.485 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:50.485 "dma_device_type": 2 00:14:50.485 } 00:14:50.485 ], 00:14:50.485 "driver_specific": {} 00:14:50.485 } 00:14:50.485 ] 00:14:50.485 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:50.485 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:14:50.485 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:50.485 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:50.485 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:14:50.485 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:50.485 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:50.485 [2024-11-19 15:20:40.737602] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:50.485 [2024-11-19 15:20:40.737698] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:50.485 [2024-11-19 15:20:40.737738] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:50.485 [2024-11-19 15:20:40.739539] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:50.485 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:50.485 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:50.485 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:50.485 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:50.485 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:50.485 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:50.485 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:50.485 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:50.485 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:50.485 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:50.485 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:50.485 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:50.485 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:50.485 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:50.485 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:50.485 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:50.485 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:50.485 "name": "Existed_Raid", 00:14:50.485 "uuid": "23a31e9f-58a9-4047-9263-e28687849ba2", 00:14:50.485 "strip_size_kb": 64, 00:14:50.485 "state": "configuring", 00:14:50.485 "raid_level": "raid5f", 00:14:50.485 "superblock": true, 00:14:50.485 "num_base_bdevs": 3, 00:14:50.485 "num_base_bdevs_discovered": 2, 00:14:50.485 "num_base_bdevs_operational": 3, 00:14:50.485 "base_bdevs_list": [ 00:14:50.485 { 00:14:50.485 "name": "BaseBdev1", 00:14:50.485 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:50.485 "is_configured": false, 00:14:50.485 "data_offset": 0, 00:14:50.485 "data_size": 0 00:14:50.485 }, 00:14:50.485 { 00:14:50.485 "name": "BaseBdev2", 00:14:50.485 "uuid": "f24d9346-c89f-4f7e-a1f2-9743f7d7f66f", 00:14:50.485 "is_configured": true, 00:14:50.485 "data_offset": 2048, 00:14:50.485 "data_size": 63488 00:14:50.485 }, 00:14:50.485 { 00:14:50.485 "name": "BaseBdev3", 00:14:50.485 "uuid": "6e089d12-83f5-4d78-ba3b-e672cda93bf0", 00:14:50.485 "is_configured": true, 00:14:50.485 "data_offset": 2048, 00:14:50.485 "data_size": 63488 00:14:50.485 } 00:14:50.485 ] 00:14:50.485 }' 00:14:50.485 15:20:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:50.485 15:20:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:51.056 15:20:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:14:51.056 15:20:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:51.056 15:20:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:51.056 [2024-11-19 15:20:41.144896] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:14:51.056 15:20:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:51.056 15:20:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:51.056 15:20:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:51.056 15:20:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:51.056 15:20:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:51.056 15:20:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:51.056 15:20:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:51.056 15:20:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:51.056 15:20:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:51.056 15:20:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:51.056 15:20:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:51.056 15:20:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:51.056 15:20:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:51.056 15:20:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:51.056 15:20:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:51.056 15:20:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:51.056 15:20:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:51.056 "name": "Existed_Raid", 00:14:51.056 "uuid": "23a31e9f-58a9-4047-9263-e28687849ba2", 00:14:51.056 "strip_size_kb": 64, 00:14:51.056 "state": "configuring", 00:14:51.056 "raid_level": "raid5f", 00:14:51.056 "superblock": true, 00:14:51.056 "num_base_bdevs": 3, 00:14:51.056 "num_base_bdevs_discovered": 1, 00:14:51.056 "num_base_bdevs_operational": 3, 00:14:51.056 "base_bdevs_list": [ 00:14:51.056 { 00:14:51.056 "name": "BaseBdev1", 00:14:51.056 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:51.056 "is_configured": false, 00:14:51.056 "data_offset": 0, 00:14:51.056 "data_size": 0 00:14:51.056 }, 00:14:51.056 { 00:14:51.056 "name": null, 00:14:51.056 "uuid": "f24d9346-c89f-4f7e-a1f2-9743f7d7f66f", 00:14:51.056 "is_configured": false, 00:14:51.056 "data_offset": 0, 00:14:51.056 "data_size": 63488 00:14:51.056 }, 00:14:51.056 { 00:14:51.056 "name": "BaseBdev3", 00:14:51.056 "uuid": "6e089d12-83f5-4d78-ba3b-e672cda93bf0", 00:14:51.056 "is_configured": true, 00:14:51.056 "data_offset": 2048, 00:14:51.056 "data_size": 63488 00:14:51.056 } 00:14:51.056 ] 00:14:51.056 }' 00:14:51.056 15:20:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:51.056 15:20:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:51.317 15:20:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:14:51.317 15:20:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:51.317 15:20:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:51.317 15:20:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:51.317 15:20:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:51.317 15:20:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:14:51.317 15:20:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:14:51.317 15:20:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:51.317 15:20:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:51.577 [2024-11-19 15:20:41.663107] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:51.577 BaseBdev1 00:14:51.577 15:20:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:51.577 15:20:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:14:51.577 15:20:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:14:51.577 15:20:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:51.577 15:20:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:14:51.577 15:20:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:51.577 15:20:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:51.577 15:20:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:51.577 15:20:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:51.578 15:20:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:51.578 15:20:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:51.578 15:20:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:14:51.578 15:20:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:51.578 15:20:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:51.578 [ 00:14:51.578 { 00:14:51.578 "name": "BaseBdev1", 00:14:51.578 "aliases": [ 00:14:51.578 "eafef5e2-ff75-4b4f-827f-127ec327247b" 00:14:51.578 ], 00:14:51.578 "product_name": "Malloc disk", 00:14:51.578 "block_size": 512, 00:14:51.578 "num_blocks": 65536, 00:14:51.578 "uuid": "eafef5e2-ff75-4b4f-827f-127ec327247b", 00:14:51.578 "assigned_rate_limits": { 00:14:51.578 "rw_ios_per_sec": 0, 00:14:51.578 "rw_mbytes_per_sec": 0, 00:14:51.578 "r_mbytes_per_sec": 0, 00:14:51.578 "w_mbytes_per_sec": 0 00:14:51.578 }, 00:14:51.578 "claimed": true, 00:14:51.578 "claim_type": "exclusive_write", 00:14:51.578 "zoned": false, 00:14:51.578 "supported_io_types": { 00:14:51.578 "read": true, 00:14:51.578 "write": true, 00:14:51.578 "unmap": true, 00:14:51.578 "flush": true, 00:14:51.578 "reset": true, 00:14:51.578 "nvme_admin": false, 00:14:51.578 "nvme_io": false, 00:14:51.578 "nvme_io_md": false, 00:14:51.578 "write_zeroes": true, 00:14:51.578 "zcopy": true, 00:14:51.578 "get_zone_info": false, 00:14:51.578 "zone_management": false, 00:14:51.578 "zone_append": false, 00:14:51.578 "compare": false, 00:14:51.578 "compare_and_write": false, 00:14:51.578 "abort": true, 00:14:51.578 "seek_hole": false, 00:14:51.578 "seek_data": false, 00:14:51.578 "copy": true, 00:14:51.578 "nvme_iov_md": false 00:14:51.578 }, 00:14:51.578 "memory_domains": [ 00:14:51.578 { 00:14:51.578 "dma_device_id": "system", 00:14:51.578 "dma_device_type": 1 00:14:51.578 }, 00:14:51.578 { 00:14:51.578 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:51.578 "dma_device_type": 2 00:14:51.578 } 00:14:51.578 ], 00:14:51.578 "driver_specific": {} 00:14:51.578 } 00:14:51.578 ] 00:14:51.578 15:20:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:51.578 15:20:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:14:51.578 15:20:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:51.578 15:20:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:51.578 15:20:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:51.578 15:20:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:51.578 15:20:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:51.578 15:20:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:51.578 15:20:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:51.578 15:20:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:51.578 15:20:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:51.578 15:20:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:51.578 15:20:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:51.578 15:20:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:51.578 15:20:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:51.578 15:20:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:51.578 15:20:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:51.578 15:20:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:51.578 "name": "Existed_Raid", 00:14:51.578 "uuid": "23a31e9f-58a9-4047-9263-e28687849ba2", 00:14:51.578 "strip_size_kb": 64, 00:14:51.578 "state": "configuring", 00:14:51.578 "raid_level": "raid5f", 00:14:51.578 "superblock": true, 00:14:51.578 "num_base_bdevs": 3, 00:14:51.578 "num_base_bdevs_discovered": 2, 00:14:51.578 "num_base_bdevs_operational": 3, 00:14:51.578 "base_bdevs_list": [ 00:14:51.578 { 00:14:51.578 "name": "BaseBdev1", 00:14:51.578 "uuid": "eafef5e2-ff75-4b4f-827f-127ec327247b", 00:14:51.578 "is_configured": true, 00:14:51.578 "data_offset": 2048, 00:14:51.578 "data_size": 63488 00:14:51.578 }, 00:14:51.578 { 00:14:51.578 "name": null, 00:14:51.578 "uuid": "f24d9346-c89f-4f7e-a1f2-9743f7d7f66f", 00:14:51.578 "is_configured": false, 00:14:51.578 "data_offset": 0, 00:14:51.578 "data_size": 63488 00:14:51.578 }, 00:14:51.578 { 00:14:51.578 "name": "BaseBdev3", 00:14:51.578 "uuid": "6e089d12-83f5-4d78-ba3b-e672cda93bf0", 00:14:51.578 "is_configured": true, 00:14:51.578 "data_offset": 2048, 00:14:51.578 "data_size": 63488 00:14:51.578 } 00:14:51.578 ] 00:14:51.578 }' 00:14:51.578 15:20:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:51.578 15:20:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:51.839 15:20:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:51.839 15:20:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:14:51.839 15:20:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:51.839 15:20:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:52.099 15:20:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:52.099 15:20:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:14:52.099 15:20:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:14:52.099 15:20:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:52.099 15:20:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:52.099 [2024-11-19 15:20:42.218171] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:14:52.099 15:20:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:52.099 15:20:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:52.099 15:20:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:52.099 15:20:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:52.099 15:20:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:52.099 15:20:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:52.099 15:20:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:52.099 15:20:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:52.099 15:20:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:52.099 15:20:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:52.099 15:20:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:52.099 15:20:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:52.099 15:20:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:52.099 15:20:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:52.100 15:20:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:52.100 15:20:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:52.100 15:20:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:52.100 "name": "Existed_Raid", 00:14:52.100 "uuid": "23a31e9f-58a9-4047-9263-e28687849ba2", 00:14:52.100 "strip_size_kb": 64, 00:14:52.100 "state": "configuring", 00:14:52.100 "raid_level": "raid5f", 00:14:52.100 "superblock": true, 00:14:52.100 "num_base_bdevs": 3, 00:14:52.100 "num_base_bdevs_discovered": 1, 00:14:52.100 "num_base_bdevs_operational": 3, 00:14:52.100 "base_bdevs_list": [ 00:14:52.100 { 00:14:52.100 "name": "BaseBdev1", 00:14:52.100 "uuid": "eafef5e2-ff75-4b4f-827f-127ec327247b", 00:14:52.100 "is_configured": true, 00:14:52.100 "data_offset": 2048, 00:14:52.100 "data_size": 63488 00:14:52.100 }, 00:14:52.100 { 00:14:52.100 "name": null, 00:14:52.100 "uuid": "f24d9346-c89f-4f7e-a1f2-9743f7d7f66f", 00:14:52.100 "is_configured": false, 00:14:52.100 "data_offset": 0, 00:14:52.100 "data_size": 63488 00:14:52.100 }, 00:14:52.100 { 00:14:52.100 "name": null, 00:14:52.100 "uuid": "6e089d12-83f5-4d78-ba3b-e672cda93bf0", 00:14:52.100 "is_configured": false, 00:14:52.100 "data_offset": 0, 00:14:52.100 "data_size": 63488 00:14:52.100 } 00:14:52.100 ] 00:14:52.100 }' 00:14:52.100 15:20:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:52.100 15:20:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:52.359 15:20:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:14:52.359 15:20:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:52.359 15:20:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:52.359 15:20:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:52.359 15:20:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:52.618 15:20:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:14:52.618 15:20:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:14:52.618 15:20:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:52.618 15:20:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:52.618 [2024-11-19 15:20:42.705330] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:52.618 15:20:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:52.618 15:20:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:52.618 15:20:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:52.618 15:20:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:52.618 15:20:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:52.618 15:20:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:52.618 15:20:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:52.618 15:20:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:52.618 15:20:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:52.618 15:20:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:52.618 15:20:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:52.618 15:20:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:52.618 15:20:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:52.618 15:20:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:52.618 15:20:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:52.618 15:20:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:52.618 15:20:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:52.618 "name": "Existed_Raid", 00:14:52.618 "uuid": "23a31e9f-58a9-4047-9263-e28687849ba2", 00:14:52.618 "strip_size_kb": 64, 00:14:52.618 "state": "configuring", 00:14:52.618 "raid_level": "raid5f", 00:14:52.618 "superblock": true, 00:14:52.618 "num_base_bdevs": 3, 00:14:52.618 "num_base_bdevs_discovered": 2, 00:14:52.618 "num_base_bdevs_operational": 3, 00:14:52.618 "base_bdevs_list": [ 00:14:52.618 { 00:14:52.618 "name": "BaseBdev1", 00:14:52.618 "uuid": "eafef5e2-ff75-4b4f-827f-127ec327247b", 00:14:52.618 "is_configured": true, 00:14:52.618 "data_offset": 2048, 00:14:52.618 "data_size": 63488 00:14:52.618 }, 00:14:52.618 { 00:14:52.618 "name": null, 00:14:52.618 "uuid": "f24d9346-c89f-4f7e-a1f2-9743f7d7f66f", 00:14:52.618 "is_configured": false, 00:14:52.618 "data_offset": 0, 00:14:52.618 "data_size": 63488 00:14:52.618 }, 00:14:52.618 { 00:14:52.618 "name": "BaseBdev3", 00:14:52.618 "uuid": "6e089d12-83f5-4d78-ba3b-e672cda93bf0", 00:14:52.618 "is_configured": true, 00:14:52.618 "data_offset": 2048, 00:14:52.618 "data_size": 63488 00:14:52.618 } 00:14:52.618 ] 00:14:52.618 }' 00:14:52.618 15:20:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:52.618 15:20:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:52.878 15:20:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:52.878 15:20:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:14:52.878 15:20:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:52.878 15:20:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:52.878 15:20:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:52.878 15:20:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:14:52.878 15:20:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:14:52.878 15:20:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:52.878 15:20:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:52.878 [2024-11-19 15:20:43.132636] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:52.878 15:20:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:52.878 15:20:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:52.878 15:20:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:52.878 15:20:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:52.878 15:20:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:52.878 15:20:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:52.878 15:20:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:52.879 15:20:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:52.879 15:20:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:52.879 15:20:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:52.879 15:20:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:52.879 15:20:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:52.879 15:20:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:52.879 15:20:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:52.879 15:20:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:52.879 15:20:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:52.879 15:20:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:52.879 "name": "Existed_Raid", 00:14:52.879 "uuid": "23a31e9f-58a9-4047-9263-e28687849ba2", 00:14:52.879 "strip_size_kb": 64, 00:14:52.879 "state": "configuring", 00:14:52.879 "raid_level": "raid5f", 00:14:52.879 "superblock": true, 00:14:52.879 "num_base_bdevs": 3, 00:14:52.879 "num_base_bdevs_discovered": 1, 00:14:52.879 "num_base_bdevs_operational": 3, 00:14:52.879 "base_bdevs_list": [ 00:14:52.879 { 00:14:52.879 "name": null, 00:14:52.879 "uuid": "eafef5e2-ff75-4b4f-827f-127ec327247b", 00:14:52.879 "is_configured": false, 00:14:52.879 "data_offset": 0, 00:14:52.879 "data_size": 63488 00:14:52.879 }, 00:14:52.879 { 00:14:52.879 "name": null, 00:14:52.879 "uuid": "f24d9346-c89f-4f7e-a1f2-9743f7d7f66f", 00:14:52.879 "is_configured": false, 00:14:52.879 "data_offset": 0, 00:14:52.879 "data_size": 63488 00:14:52.879 }, 00:14:52.879 { 00:14:52.879 "name": "BaseBdev3", 00:14:52.879 "uuid": "6e089d12-83f5-4d78-ba3b-e672cda93bf0", 00:14:52.879 "is_configured": true, 00:14:52.879 "data_offset": 2048, 00:14:52.879 "data_size": 63488 00:14:52.879 } 00:14:52.879 ] 00:14:52.879 }' 00:14:52.879 15:20:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:52.879 15:20:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:53.448 15:20:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:53.448 15:20:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:14:53.448 15:20:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:53.448 15:20:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:53.449 15:20:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:53.449 15:20:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:14:53.449 15:20:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:14:53.449 15:20:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:53.449 15:20:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:53.449 [2024-11-19 15:20:43.610354] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:53.449 15:20:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:53.449 15:20:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:53.449 15:20:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:53.449 15:20:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:53.449 15:20:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:53.449 15:20:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:53.449 15:20:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:53.449 15:20:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:53.449 15:20:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:53.449 15:20:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:53.449 15:20:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:53.449 15:20:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:53.449 15:20:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:53.449 15:20:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:53.449 15:20:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:53.449 15:20:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:53.449 15:20:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:53.449 "name": "Existed_Raid", 00:14:53.449 "uuid": "23a31e9f-58a9-4047-9263-e28687849ba2", 00:14:53.449 "strip_size_kb": 64, 00:14:53.449 "state": "configuring", 00:14:53.449 "raid_level": "raid5f", 00:14:53.449 "superblock": true, 00:14:53.449 "num_base_bdevs": 3, 00:14:53.449 "num_base_bdevs_discovered": 2, 00:14:53.449 "num_base_bdevs_operational": 3, 00:14:53.449 "base_bdevs_list": [ 00:14:53.449 { 00:14:53.449 "name": null, 00:14:53.449 "uuid": "eafef5e2-ff75-4b4f-827f-127ec327247b", 00:14:53.449 "is_configured": false, 00:14:53.449 "data_offset": 0, 00:14:53.449 "data_size": 63488 00:14:53.449 }, 00:14:53.449 { 00:14:53.449 "name": "BaseBdev2", 00:14:53.449 "uuid": "f24d9346-c89f-4f7e-a1f2-9743f7d7f66f", 00:14:53.449 "is_configured": true, 00:14:53.449 "data_offset": 2048, 00:14:53.449 "data_size": 63488 00:14:53.449 }, 00:14:53.449 { 00:14:53.449 "name": "BaseBdev3", 00:14:53.449 "uuid": "6e089d12-83f5-4d78-ba3b-e672cda93bf0", 00:14:53.449 "is_configured": true, 00:14:53.449 "data_offset": 2048, 00:14:53.449 "data_size": 63488 00:14:53.449 } 00:14:53.449 ] 00:14:53.449 }' 00:14:53.449 15:20:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:53.449 15:20:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:54.018 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:14:54.018 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:54.018 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:54.018 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:54.018 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:54.018 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:14:54.018 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:54.018 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:14:54.018 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:54.018 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:54.018 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:54.018 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u eafef5e2-ff75-4b4f-827f-127ec327247b 00:14:54.018 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:54.018 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:54.018 [2024-11-19 15:20:44.188451] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:14:54.018 [2024-11-19 15:20:44.188694] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:14:54.018 [2024-11-19 15:20:44.188715] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:14:54.018 [2024-11-19 15:20:44.188966] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:14:54.018 NewBaseBdev 00:14:54.018 [2024-11-19 15:20:44.189356] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:14:54.018 [2024-11-19 15:20:44.189377] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:14:54.018 [2024-11-19 15:20:44.189480] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:54.018 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:54.018 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:14:54.018 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:14:54.018 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:54.018 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:14:54.018 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:54.018 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:54.018 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:54.018 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:54.018 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:54.018 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:54.018 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:14:54.018 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:54.018 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:54.018 [ 00:14:54.018 { 00:14:54.018 "name": "NewBaseBdev", 00:14:54.018 "aliases": [ 00:14:54.018 "eafef5e2-ff75-4b4f-827f-127ec327247b" 00:14:54.018 ], 00:14:54.018 "product_name": "Malloc disk", 00:14:54.018 "block_size": 512, 00:14:54.018 "num_blocks": 65536, 00:14:54.018 "uuid": "eafef5e2-ff75-4b4f-827f-127ec327247b", 00:14:54.018 "assigned_rate_limits": { 00:14:54.018 "rw_ios_per_sec": 0, 00:14:54.018 "rw_mbytes_per_sec": 0, 00:14:54.018 "r_mbytes_per_sec": 0, 00:14:54.018 "w_mbytes_per_sec": 0 00:14:54.018 }, 00:14:54.018 "claimed": true, 00:14:54.018 "claim_type": "exclusive_write", 00:14:54.018 "zoned": false, 00:14:54.018 "supported_io_types": { 00:14:54.018 "read": true, 00:14:54.018 "write": true, 00:14:54.018 "unmap": true, 00:14:54.018 "flush": true, 00:14:54.018 "reset": true, 00:14:54.018 "nvme_admin": false, 00:14:54.018 "nvme_io": false, 00:14:54.018 "nvme_io_md": false, 00:14:54.018 "write_zeroes": true, 00:14:54.018 "zcopy": true, 00:14:54.018 "get_zone_info": false, 00:14:54.018 "zone_management": false, 00:14:54.018 "zone_append": false, 00:14:54.018 "compare": false, 00:14:54.018 "compare_and_write": false, 00:14:54.018 "abort": true, 00:14:54.018 "seek_hole": false, 00:14:54.018 "seek_data": false, 00:14:54.018 "copy": true, 00:14:54.018 "nvme_iov_md": false 00:14:54.018 }, 00:14:54.018 "memory_domains": [ 00:14:54.018 { 00:14:54.018 "dma_device_id": "system", 00:14:54.018 "dma_device_type": 1 00:14:54.018 }, 00:14:54.018 { 00:14:54.018 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:54.018 "dma_device_type": 2 00:14:54.018 } 00:14:54.018 ], 00:14:54.018 "driver_specific": {} 00:14:54.018 } 00:14:54.018 ] 00:14:54.018 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:54.018 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:14:54.018 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:14:54.018 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:54.018 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:54.018 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:54.018 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:54.018 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:54.018 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:54.018 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:54.018 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:54.018 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:54.018 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:54.018 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:54.018 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:54.018 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:54.019 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:54.019 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:54.019 "name": "Existed_Raid", 00:14:54.019 "uuid": "23a31e9f-58a9-4047-9263-e28687849ba2", 00:14:54.019 "strip_size_kb": 64, 00:14:54.019 "state": "online", 00:14:54.019 "raid_level": "raid5f", 00:14:54.019 "superblock": true, 00:14:54.019 "num_base_bdevs": 3, 00:14:54.019 "num_base_bdevs_discovered": 3, 00:14:54.019 "num_base_bdevs_operational": 3, 00:14:54.019 "base_bdevs_list": [ 00:14:54.019 { 00:14:54.019 "name": "NewBaseBdev", 00:14:54.019 "uuid": "eafef5e2-ff75-4b4f-827f-127ec327247b", 00:14:54.019 "is_configured": true, 00:14:54.019 "data_offset": 2048, 00:14:54.019 "data_size": 63488 00:14:54.019 }, 00:14:54.019 { 00:14:54.019 "name": "BaseBdev2", 00:14:54.019 "uuid": "f24d9346-c89f-4f7e-a1f2-9743f7d7f66f", 00:14:54.019 "is_configured": true, 00:14:54.019 "data_offset": 2048, 00:14:54.019 "data_size": 63488 00:14:54.019 }, 00:14:54.019 { 00:14:54.019 "name": "BaseBdev3", 00:14:54.019 "uuid": "6e089d12-83f5-4d78-ba3b-e672cda93bf0", 00:14:54.019 "is_configured": true, 00:14:54.019 "data_offset": 2048, 00:14:54.019 "data_size": 63488 00:14:54.019 } 00:14:54.019 ] 00:14:54.019 }' 00:14:54.019 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:54.019 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:54.588 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:14:54.588 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:14:54.588 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:54.588 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:54.588 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:14:54.588 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:54.588 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:14:54.588 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:54.588 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:54.588 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:54.588 [2024-11-19 15:20:44.663943] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:54.588 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:54.588 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:54.588 "name": "Existed_Raid", 00:14:54.588 "aliases": [ 00:14:54.588 "23a31e9f-58a9-4047-9263-e28687849ba2" 00:14:54.588 ], 00:14:54.588 "product_name": "Raid Volume", 00:14:54.588 "block_size": 512, 00:14:54.588 "num_blocks": 126976, 00:14:54.588 "uuid": "23a31e9f-58a9-4047-9263-e28687849ba2", 00:14:54.588 "assigned_rate_limits": { 00:14:54.588 "rw_ios_per_sec": 0, 00:14:54.588 "rw_mbytes_per_sec": 0, 00:14:54.588 "r_mbytes_per_sec": 0, 00:14:54.588 "w_mbytes_per_sec": 0 00:14:54.588 }, 00:14:54.588 "claimed": false, 00:14:54.588 "zoned": false, 00:14:54.588 "supported_io_types": { 00:14:54.588 "read": true, 00:14:54.588 "write": true, 00:14:54.588 "unmap": false, 00:14:54.588 "flush": false, 00:14:54.588 "reset": true, 00:14:54.588 "nvme_admin": false, 00:14:54.588 "nvme_io": false, 00:14:54.588 "nvme_io_md": false, 00:14:54.588 "write_zeroes": true, 00:14:54.588 "zcopy": false, 00:14:54.588 "get_zone_info": false, 00:14:54.588 "zone_management": false, 00:14:54.588 "zone_append": false, 00:14:54.588 "compare": false, 00:14:54.588 "compare_and_write": false, 00:14:54.588 "abort": false, 00:14:54.588 "seek_hole": false, 00:14:54.588 "seek_data": false, 00:14:54.588 "copy": false, 00:14:54.588 "nvme_iov_md": false 00:14:54.588 }, 00:14:54.588 "driver_specific": { 00:14:54.588 "raid": { 00:14:54.588 "uuid": "23a31e9f-58a9-4047-9263-e28687849ba2", 00:14:54.588 "strip_size_kb": 64, 00:14:54.588 "state": "online", 00:14:54.588 "raid_level": "raid5f", 00:14:54.588 "superblock": true, 00:14:54.588 "num_base_bdevs": 3, 00:14:54.588 "num_base_bdevs_discovered": 3, 00:14:54.588 "num_base_bdevs_operational": 3, 00:14:54.588 "base_bdevs_list": [ 00:14:54.588 { 00:14:54.588 "name": "NewBaseBdev", 00:14:54.588 "uuid": "eafef5e2-ff75-4b4f-827f-127ec327247b", 00:14:54.588 "is_configured": true, 00:14:54.588 "data_offset": 2048, 00:14:54.588 "data_size": 63488 00:14:54.588 }, 00:14:54.588 { 00:14:54.588 "name": "BaseBdev2", 00:14:54.588 "uuid": "f24d9346-c89f-4f7e-a1f2-9743f7d7f66f", 00:14:54.588 "is_configured": true, 00:14:54.588 "data_offset": 2048, 00:14:54.588 "data_size": 63488 00:14:54.588 }, 00:14:54.588 { 00:14:54.588 "name": "BaseBdev3", 00:14:54.588 "uuid": "6e089d12-83f5-4d78-ba3b-e672cda93bf0", 00:14:54.588 "is_configured": true, 00:14:54.588 "data_offset": 2048, 00:14:54.588 "data_size": 63488 00:14:54.588 } 00:14:54.588 ] 00:14:54.588 } 00:14:54.588 } 00:14:54.588 }' 00:14:54.588 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:54.588 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:14:54.588 BaseBdev2 00:14:54.588 BaseBdev3' 00:14:54.588 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:54.588 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:54.588 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:54.588 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:14:54.588 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:54.588 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:54.588 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:54.588 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:54.588 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:54.588 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:54.588 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:54.588 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:54.588 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:14:54.588 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:54.588 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:54.588 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:54.588 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:54.588 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:54.588 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:54.588 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:14:54.588 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:54.588 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:54.588 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:54.588 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:54.848 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:54.848 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:54.848 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:54.848 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:54.848 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:54.849 [2024-11-19 15:20:44.939265] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:54.849 [2024-11-19 15:20:44.939287] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:54.849 [2024-11-19 15:20:44.939346] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:54.849 [2024-11-19 15:20:44.939567] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:54.849 [2024-11-19 15:20:44.939579] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:14:54.849 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:54.849 15:20:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 91067 00:14:54.849 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 91067 ']' 00:14:54.849 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 91067 00:14:54.849 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:14:54.849 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:14:54.849 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 91067 00:14:54.849 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:14:54.849 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:14:54.849 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 91067' 00:14:54.849 killing process with pid 91067 00:14:54.849 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 91067 00:14:54.849 [2024-11-19 15:20:44.991997] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:54.849 15:20:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 91067 00:14:54.849 [2024-11-19 15:20:45.022873] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:55.109 15:20:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:14:55.109 00:14:55.109 real 0m8.863s 00:14:55.109 user 0m15.098s 00:14:55.109 sys 0m1.987s 00:14:55.109 15:20:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:14:55.109 ************************************ 00:14:55.109 END TEST raid5f_state_function_test_sb 00:14:55.109 ************************************ 00:14:55.109 15:20:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:55.110 15:20:45 bdev_raid -- bdev/bdev_raid.sh@988 -- # run_test raid5f_superblock_test raid_superblock_test raid5f 3 00:14:55.110 15:20:45 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:14:55.110 15:20:45 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:14:55.110 15:20:45 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:55.110 ************************************ 00:14:55.110 START TEST raid5f_superblock_test 00:14:55.110 ************************************ 00:14:55.110 15:20:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid5f 3 00:14:55.110 15:20:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid5f 00:14:55.110 15:20:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:14:55.110 15:20:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:14:55.110 15:20:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:14:55.110 15:20:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:14:55.110 15:20:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:14:55.110 15:20:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:14:55.110 15:20:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:14:55.110 15:20:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:14:55.110 15:20:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:14:55.110 15:20:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:14:55.110 15:20:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:14:55.110 15:20:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:14:55.110 15:20:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid5f '!=' raid1 ']' 00:14:55.110 15:20:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:14:55.110 15:20:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:14:55.110 15:20:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=91671 00:14:55.110 15:20:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:14:55.110 15:20:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 91671 00:14:55.110 15:20:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 91671 ']' 00:14:55.110 15:20:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:55.110 15:20:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:14:55.110 15:20:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:55.110 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:55.110 15:20:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:14:55.110 15:20:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:55.110 [2024-11-19 15:20:45.413709] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:14:55.110 [2024-11-19 15:20:45.414466] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid91671 ] 00:14:55.369 [2024-11-19 15:20:45.572040] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:55.369 [2024-11-19 15:20:45.597495] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:14:55.369 [2024-11-19 15:20:45.640546] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:55.369 [2024-11-19 15:20:45.640661] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:55.939 15:20:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:14:55.939 15:20:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:14:55.939 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:14:55.939 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:14:55.939 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:14:55.939 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:14:55.939 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:14:55.939 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:14:55.939 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:14:55.939 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:14:55.939 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:14:55.939 15:20:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:55.939 15:20:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:55.939 malloc1 00:14:55.939 15:20:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:55.939 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:14:55.939 15:20:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:55.939 15:20:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:55.939 [2024-11-19 15:20:46.243205] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:14:55.939 [2024-11-19 15:20:46.243274] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:55.939 [2024-11-19 15:20:46.243300] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:14:55.939 [2024-11-19 15:20:46.243314] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:55.939 [2024-11-19 15:20:46.245413] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:55.939 [2024-11-19 15:20:46.245459] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:14:55.939 pt1 00:14:55.939 15:20:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:55.939 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:14:55.939 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:14:55.939 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:14:55.939 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:14:55.939 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:14:55.939 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:14:55.939 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:14:55.939 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:14:55.939 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:14:55.939 15:20:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:55.939 15:20:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:55.939 malloc2 00:14:55.939 15:20:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:55.939 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:14:55.939 15:20:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:55.939 15:20:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:55.939 [2024-11-19 15:20:46.271905] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:14:55.939 [2024-11-19 15:20:46.272039] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:55.939 [2024-11-19 15:20:46.272073] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:14:55.939 [2024-11-19 15:20:46.272103] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:55.939 [2024-11-19 15:20:46.274155] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:55.939 [2024-11-19 15:20:46.274226] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:14:55.939 pt2 00:14:56.199 15:20:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.199 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:14:56.199 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:14:56.199 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:14:56.199 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:14:56.199 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:14:56.199 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:14:56.199 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:14:56.199 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:14:56.199 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:14:56.199 15:20:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.199 15:20:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:56.199 malloc3 00:14:56.199 15:20:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.199 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:14:56.199 15:20:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.199 15:20:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:56.199 [2024-11-19 15:20:46.304431] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:14:56.199 [2024-11-19 15:20:46.304539] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:56.199 [2024-11-19 15:20:46.304591] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:14:56.199 [2024-11-19 15:20:46.304626] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:56.199 [2024-11-19 15:20:46.306673] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:56.199 [2024-11-19 15:20:46.306744] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:14:56.199 pt3 00:14:56.199 15:20:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.199 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:14:56.199 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:14:56.199 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:14:56.199 15:20:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.199 15:20:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:56.199 [2024-11-19 15:20:46.316471] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:14:56.199 [2024-11-19 15:20:46.318350] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:56.199 [2024-11-19 15:20:46.318437] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:14:56.199 [2024-11-19 15:20:46.318618] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:14:56.199 [2024-11-19 15:20:46.318686] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:14:56.199 [2024-11-19 15:20:46.318961] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:14:56.199 [2024-11-19 15:20:46.319422] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:14:56.199 [2024-11-19 15:20:46.319471] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:14:56.199 [2024-11-19 15:20:46.319610] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:56.199 15:20:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.199 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:56.199 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:56.199 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:56.199 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:56.199 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:56.199 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:56.199 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:56.199 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:56.199 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:56.199 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:56.199 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:56.199 15:20:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.199 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:56.199 15:20:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:56.199 15:20:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.199 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:56.199 "name": "raid_bdev1", 00:14:56.199 "uuid": "4f449536-f5b5-45e3-abab-fe44e561a816", 00:14:56.199 "strip_size_kb": 64, 00:14:56.199 "state": "online", 00:14:56.199 "raid_level": "raid5f", 00:14:56.199 "superblock": true, 00:14:56.199 "num_base_bdevs": 3, 00:14:56.199 "num_base_bdevs_discovered": 3, 00:14:56.199 "num_base_bdevs_operational": 3, 00:14:56.199 "base_bdevs_list": [ 00:14:56.199 { 00:14:56.199 "name": "pt1", 00:14:56.199 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:56.199 "is_configured": true, 00:14:56.199 "data_offset": 2048, 00:14:56.199 "data_size": 63488 00:14:56.199 }, 00:14:56.199 { 00:14:56.199 "name": "pt2", 00:14:56.199 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:56.199 "is_configured": true, 00:14:56.199 "data_offset": 2048, 00:14:56.199 "data_size": 63488 00:14:56.199 }, 00:14:56.199 { 00:14:56.199 "name": "pt3", 00:14:56.199 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:56.199 "is_configured": true, 00:14:56.199 "data_offset": 2048, 00:14:56.199 "data_size": 63488 00:14:56.199 } 00:14:56.199 ] 00:14:56.199 }' 00:14:56.199 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:56.199 15:20:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:56.459 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:14:56.459 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:14:56.459 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:56.459 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:56.459 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:14:56.459 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:56.459 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:56.459 15:20:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.459 15:20:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:56.459 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:56.459 [2024-11-19 15:20:46.744404] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:56.459 15:20:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.459 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:56.459 "name": "raid_bdev1", 00:14:56.459 "aliases": [ 00:14:56.459 "4f449536-f5b5-45e3-abab-fe44e561a816" 00:14:56.459 ], 00:14:56.459 "product_name": "Raid Volume", 00:14:56.459 "block_size": 512, 00:14:56.459 "num_blocks": 126976, 00:14:56.459 "uuid": "4f449536-f5b5-45e3-abab-fe44e561a816", 00:14:56.459 "assigned_rate_limits": { 00:14:56.459 "rw_ios_per_sec": 0, 00:14:56.459 "rw_mbytes_per_sec": 0, 00:14:56.459 "r_mbytes_per_sec": 0, 00:14:56.459 "w_mbytes_per_sec": 0 00:14:56.459 }, 00:14:56.459 "claimed": false, 00:14:56.459 "zoned": false, 00:14:56.459 "supported_io_types": { 00:14:56.459 "read": true, 00:14:56.459 "write": true, 00:14:56.459 "unmap": false, 00:14:56.459 "flush": false, 00:14:56.459 "reset": true, 00:14:56.459 "nvme_admin": false, 00:14:56.459 "nvme_io": false, 00:14:56.460 "nvme_io_md": false, 00:14:56.460 "write_zeroes": true, 00:14:56.460 "zcopy": false, 00:14:56.460 "get_zone_info": false, 00:14:56.460 "zone_management": false, 00:14:56.460 "zone_append": false, 00:14:56.460 "compare": false, 00:14:56.460 "compare_and_write": false, 00:14:56.460 "abort": false, 00:14:56.460 "seek_hole": false, 00:14:56.460 "seek_data": false, 00:14:56.460 "copy": false, 00:14:56.460 "nvme_iov_md": false 00:14:56.460 }, 00:14:56.460 "driver_specific": { 00:14:56.460 "raid": { 00:14:56.460 "uuid": "4f449536-f5b5-45e3-abab-fe44e561a816", 00:14:56.460 "strip_size_kb": 64, 00:14:56.460 "state": "online", 00:14:56.460 "raid_level": "raid5f", 00:14:56.460 "superblock": true, 00:14:56.460 "num_base_bdevs": 3, 00:14:56.460 "num_base_bdevs_discovered": 3, 00:14:56.460 "num_base_bdevs_operational": 3, 00:14:56.460 "base_bdevs_list": [ 00:14:56.460 { 00:14:56.460 "name": "pt1", 00:14:56.460 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:56.460 "is_configured": true, 00:14:56.460 "data_offset": 2048, 00:14:56.460 "data_size": 63488 00:14:56.460 }, 00:14:56.460 { 00:14:56.460 "name": "pt2", 00:14:56.460 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:56.460 "is_configured": true, 00:14:56.460 "data_offset": 2048, 00:14:56.460 "data_size": 63488 00:14:56.460 }, 00:14:56.460 { 00:14:56.460 "name": "pt3", 00:14:56.460 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:56.460 "is_configured": true, 00:14:56.460 "data_offset": 2048, 00:14:56.460 "data_size": 63488 00:14:56.460 } 00:14:56.460 ] 00:14:56.460 } 00:14:56.460 } 00:14:56.460 }' 00:14:56.460 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:56.720 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:14:56.720 pt2 00:14:56.720 pt3' 00:14:56.720 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:56.720 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:56.720 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:56.720 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:14:56.720 15:20:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.720 15:20:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:56.720 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:56.720 15:20:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.720 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:56.720 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:56.720 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:56.720 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:56.720 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:14:56.720 15:20:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.720 15:20:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:56.720 15:20:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.720 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:56.720 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:56.720 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:56.720 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:14:56.720 15:20:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.720 15:20:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:56.720 15:20:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:56.720 15:20:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.720 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:56.720 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:56.720 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:56.720 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:14:56.720 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.720 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:56.721 [2024-11-19 15:20:47.031994] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:56.721 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.981 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=4f449536-f5b5-45e3-abab-fe44e561a816 00:14:56.981 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 4f449536-f5b5-45e3-abab-fe44e561a816 ']' 00:14:56.981 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:56.981 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.981 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:56.981 [2024-11-19 15:20:47.079842] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:56.981 [2024-11-19 15:20:47.079906] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:56.981 [2024-11-19 15:20:47.080038] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:56.981 [2024-11-19 15:20:47.080139] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:56.981 [2024-11-19 15:20:47.080192] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:14:56.981 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.981 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:56.981 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.981 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:56.981 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:14:56.981 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.981 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:14:56.981 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:14:56.981 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:14:56.981 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:14:56.981 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.981 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:56.981 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.981 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:14:56.981 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:14:56.981 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.981 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:56.981 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.981 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:14:56.981 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:14:56.981 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.981 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:56.981 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.981 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:14:56.981 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:14:56.981 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.981 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:56.981 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.981 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:14:56.981 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:14:56.981 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:14:56.981 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:14:56.981 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:14:56.982 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:14:56.982 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:14:56.982 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:14:56.982 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:14:56.982 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.982 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:56.982 [2024-11-19 15:20:47.231607] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:14:56.982 [2024-11-19 15:20:47.233479] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:14:56.982 [2024-11-19 15:20:47.233559] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:14:56.982 [2024-11-19 15:20:47.233629] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:14:56.982 [2024-11-19 15:20:47.233718] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:14:56.982 [2024-11-19 15:20:47.233789] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:14:56.982 [2024-11-19 15:20:47.233835] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:56.982 [2024-11-19 15:20:47.233880] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:14:56.982 request: 00:14:56.982 { 00:14:56.982 "name": "raid_bdev1", 00:14:56.982 "raid_level": "raid5f", 00:14:56.982 "base_bdevs": [ 00:14:56.982 "malloc1", 00:14:56.982 "malloc2", 00:14:56.982 "malloc3" 00:14:56.982 ], 00:14:56.982 "strip_size_kb": 64, 00:14:56.982 "superblock": false, 00:14:56.982 "method": "bdev_raid_create", 00:14:56.982 "req_id": 1 00:14:56.982 } 00:14:56.982 Got JSON-RPC error response 00:14:56.982 response: 00:14:56.982 { 00:14:56.982 "code": -17, 00:14:56.982 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:14:56.982 } 00:14:56.982 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:14:56.982 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:14:56.982 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:14:56.982 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:14:56.982 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:14:56.982 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:56.982 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:14:56.982 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.982 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:56.982 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.982 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:14:56.982 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:14:56.982 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:14:56.982 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.982 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:56.982 [2024-11-19 15:20:47.299456] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:14:56.982 [2024-11-19 15:20:47.299562] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:56.982 [2024-11-19 15:20:47.299591] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:14:56.982 [2024-11-19 15:20:47.299620] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:56.982 [2024-11-19 15:20:47.301683] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:56.982 [2024-11-19 15:20:47.301773] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:14:56.982 [2024-11-19 15:20:47.301849] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:14:56.982 [2024-11-19 15:20:47.301910] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:14:56.982 pt1 00:14:56.982 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.982 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:14:56.982 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:56.982 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:56.982 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:56.982 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:56.982 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:56.982 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:56.982 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:56.982 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:56.982 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:56.982 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:56.982 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:56.982 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.982 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:57.242 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:57.242 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:57.242 "name": "raid_bdev1", 00:14:57.242 "uuid": "4f449536-f5b5-45e3-abab-fe44e561a816", 00:14:57.242 "strip_size_kb": 64, 00:14:57.242 "state": "configuring", 00:14:57.242 "raid_level": "raid5f", 00:14:57.242 "superblock": true, 00:14:57.242 "num_base_bdevs": 3, 00:14:57.242 "num_base_bdevs_discovered": 1, 00:14:57.242 "num_base_bdevs_operational": 3, 00:14:57.242 "base_bdevs_list": [ 00:14:57.242 { 00:14:57.242 "name": "pt1", 00:14:57.242 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:57.242 "is_configured": true, 00:14:57.242 "data_offset": 2048, 00:14:57.242 "data_size": 63488 00:14:57.242 }, 00:14:57.242 { 00:14:57.242 "name": null, 00:14:57.242 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:57.242 "is_configured": false, 00:14:57.242 "data_offset": 2048, 00:14:57.242 "data_size": 63488 00:14:57.242 }, 00:14:57.242 { 00:14:57.242 "name": null, 00:14:57.242 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:57.242 "is_configured": false, 00:14:57.242 "data_offset": 2048, 00:14:57.242 "data_size": 63488 00:14:57.242 } 00:14:57.242 ] 00:14:57.242 }' 00:14:57.242 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:57.242 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:57.502 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:14:57.502 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:14:57.502 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:57.502 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:57.502 [2024-11-19 15:20:47.698792] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:14:57.502 [2024-11-19 15:20:47.698910] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:57.502 [2024-11-19 15:20:47.698947] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:14:57.502 [2024-11-19 15:20:47.698986] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:57.502 [2024-11-19 15:20:47.699371] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:57.502 [2024-11-19 15:20:47.699430] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:14:57.502 [2024-11-19 15:20:47.699517] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:14:57.502 [2024-11-19 15:20:47.699566] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:57.502 pt2 00:14:57.502 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:57.502 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:14:57.502 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:57.502 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:57.502 [2024-11-19 15:20:47.710778] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:14:57.502 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:57.502 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:14:57.502 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:57.502 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:57.502 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:57.502 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:57.502 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:57.502 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:57.502 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:57.502 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:57.502 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:57.502 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:57.502 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:57.502 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:57.502 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:57.502 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:57.502 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:57.502 "name": "raid_bdev1", 00:14:57.502 "uuid": "4f449536-f5b5-45e3-abab-fe44e561a816", 00:14:57.502 "strip_size_kb": 64, 00:14:57.502 "state": "configuring", 00:14:57.502 "raid_level": "raid5f", 00:14:57.502 "superblock": true, 00:14:57.502 "num_base_bdevs": 3, 00:14:57.502 "num_base_bdevs_discovered": 1, 00:14:57.502 "num_base_bdevs_operational": 3, 00:14:57.502 "base_bdevs_list": [ 00:14:57.502 { 00:14:57.502 "name": "pt1", 00:14:57.502 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:57.502 "is_configured": true, 00:14:57.502 "data_offset": 2048, 00:14:57.502 "data_size": 63488 00:14:57.502 }, 00:14:57.502 { 00:14:57.502 "name": null, 00:14:57.502 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:57.502 "is_configured": false, 00:14:57.502 "data_offset": 0, 00:14:57.502 "data_size": 63488 00:14:57.502 }, 00:14:57.502 { 00:14:57.502 "name": null, 00:14:57.502 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:57.502 "is_configured": false, 00:14:57.502 "data_offset": 2048, 00:14:57.502 "data_size": 63488 00:14:57.502 } 00:14:57.502 ] 00:14:57.502 }' 00:14:57.502 15:20:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:57.502 15:20:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:57.761 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:14:57.761 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:14:57.761 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:14:57.761 15:20:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:57.761 15:20:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:57.761 [2024-11-19 15:20:48.090082] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:14:57.761 [2024-11-19 15:20:48.090168] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:57.761 [2024-11-19 15:20:48.090199] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:14:57.761 [2024-11-19 15:20:48.090225] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:57.761 [2024-11-19 15:20:48.090583] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:57.761 [2024-11-19 15:20:48.090636] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:14:57.761 [2024-11-19 15:20:48.090713] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:14:57.761 [2024-11-19 15:20:48.090755] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:57.761 pt2 00:14:57.761 15:20:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:57.761 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:14:57.761 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:14:57.761 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:14:57.761 15:20:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:57.761 15:20:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:58.021 [2024-11-19 15:20:48.102077] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:14:58.021 [2024-11-19 15:20:48.102157] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:58.021 [2024-11-19 15:20:48.102188] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:14:58.021 [2024-11-19 15:20:48.102213] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:58.021 [2024-11-19 15:20:48.102538] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:58.021 [2024-11-19 15:20:48.102591] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:14:58.021 [2024-11-19 15:20:48.102665] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:14:58.021 [2024-11-19 15:20:48.102706] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:14:58.021 [2024-11-19 15:20:48.102808] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:14:58.022 [2024-11-19 15:20:48.102858] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:14:58.022 [2024-11-19 15:20:48.103102] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:14:58.022 [2024-11-19 15:20:48.103488] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:14:58.022 [2024-11-19 15:20:48.103534] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:14:58.022 [2024-11-19 15:20:48.103660] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:58.022 pt3 00:14:58.022 15:20:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:58.022 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:14:58.022 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:14:58.022 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:58.022 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:58.022 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:58.022 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:58.022 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:58.022 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:58.022 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:58.022 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:58.022 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:58.022 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:58.022 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:58.022 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:58.022 15:20:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:58.022 15:20:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:58.022 15:20:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:58.022 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:58.022 "name": "raid_bdev1", 00:14:58.022 "uuid": "4f449536-f5b5-45e3-abab-fe44e561a816", 00:14:58.022 "strip_size_kb": 64, 00:14:58.022 "state": "online", 00:14:58.022 "raid_level": "raid5f", 00:14:58.022 "superblock": true, 00:14:58.022 "num_base_bdevs": 3, 00:14:58.022 "num_base_bdevs_discovered": 3, 00:14:58.022 "num_base_bdevs_operational": 3, 00:14:58.022 "base_bdevs_list": [ 00:14:58.022 { 00:14:58.022 "name": "pt1", 00:14:58.022 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:58.022 "is_configured": true, 00:14:58.022 "data_offset": 2048, 00:14:58.022 "data_size": 63488 00:14:58.022 }, 00:14:58.022 { 00:14:58.022 "name": "pt2", 00:14:58.022 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:58.022 "is_configured": true, 00:14:58.022 "data_offset": 2048, 00:14:58.022 "data_size": 63488 00:14:58.022 }, 00:14:58.022 { 00:14:58.022 "name": "pt3", 00:14:58.022 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:58.022 "is_configured": true, 00:14:58.022 "data_offset": 2048, 00:14:58.022 "data_size": 63488 00:14:58.022 } 00:14:58.022 ] 00:14:58.022 }' 00:14:58.022 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:58.022 15:20:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:58.282 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:14:58.282 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:14:58.282 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:58.282 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:58.282 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:14:58.282 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:58.282 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:58.282 15:20:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:58.282 15:20:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:58.282 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:58.282 [2024-11-19 15:20:48.581398] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:58.282 15:20:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:58.542 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:58.542 "name": "raid_bdev1", 00:14:58.542 "aliases": [ 00:14:58.542 "4f449536-f5b5-45e3-abab-fe44e561a816" 00:14:58.542 ], 00:14:58.542 "product_name": "Raid Volume", 00:14:58.542 "block_size": 512, 00:14:58.542 "num_blocks": 126976, 00:14:58.542 "uuid": "4f449536-f5b5-45e3-abab-fe44e561a816", 00:14:58.542 "assigned_rate_limits": { 00:14:58.542 "rw_ios_per_sec": 0, 00:14:58.542 "rw_mbytes_per_sec": 0, 00:14:58.542 "r_mbytes_per_sec": 0, 00:14:58.542 "w_mbytes_per_sec": 0 00:14:58.542 }, 00:14:58.542 "claimed": false, 00:14:58.542 "zoned": false, 00:14:58.542 "supported_io_types": { 00:14:58.542 "read": true, 00:14:58.542 "write": true, 00:14:58.542 "unmap": false, 00:14:58.542 "flush": false, 00:14:58.543 "reset": true, 00:14:58.543 "nvme_admin": false, 00:14:58.543 "nvme_io": false, 00:14:58.543 "nvme_io_md": false, 00:14:58.543 "write_zeroes": true, 00:14:58.543 "zcopy": false, 00:14:58.543 "get_zone_info": false, 00:14:58.543 "zone_management": false, 00:14:58.543 "zone_append": false, 00:14:58.543 "compare": false, 00:14:58.543 "compare_and_write": false, 00:14:58.543 "abort": false, 00:14:58.543 "seek_hole": false, 00:14:58.543 "seek_data": false, 00:14:58.543 "copy": false, 00:14:58.543 "nvme_iov_md": false 00:14:58.543 }, 00:14:58.543 "driver_specific": { 00:14:58.543 "raid": { 00:14:58.543 "uuid": "4f449536-f5b5-45e3-abab-fe44e561a816", 00:14:58.543 "strip_size_kb": 64, 00:14:58.543 "state": "online", 00:14:58.543 "raid_level": "raid5f", 00:14:58.543 "superblock": true, 00:14:58.543 "num_base_bdevs": 3, 00:14:58.543 "num_base_bdevs_discovered": 3, 00:14:58.543 "num_base_bdevs_operational": 3, 00:14:58.543 "base_bdevs_list": [ 00:14:58.543 { 00:14:58.543 "name": "pt1", 00:14:58.543 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:58.543 "is_configured": true, 00:14:58.543 "data_offset": 2048, 00:14:58.543 "data_size": 63488 00:14:58.543 }, 00:14:58.543 { 00:14:58.543 "name": "pt2", 00:14:58.543 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:58.543 "is_configured": true, 00:14:58.543 "data_offset": 2048, 00:14:58.543 "data_size": 63488 00:14:58.543 }, 00:14:58.543 { 00:14:58.543 "name": "pt3", 00:14:58.543 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:58.543 "is_configured": true, 00:14:58.543 "data_offset": 2048, 00:14:58.543 "data_size": 63488 00:14:58.543 } 00:14:58.543 ] 00:14:58.543 } 00:14:58.543 } 00:14:58.543 }' 00:14:58.543 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:58.543 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:14:58.543 pt2 00:14:58.543 pt3' 00:14:58.543 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:58.543 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:58.543 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:58.543 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:14:58.543 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:58.543 15:20:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:58.543 15:20:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:58.543 15:20:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:58.543 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:58.543 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:58.543 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:58.543 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:14:58.543 15:20:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:58.543 15:20:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:58.543 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:58.543 15:20:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:58.543 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:58.543 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:58.543 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:58.543 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:14:58.543 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:58.543 15:20:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:58.543 15:20:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:58.543 15:20:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:58.543 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:58.543 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:58.543 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:14:58.543 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:58.543 15:20:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:58.543 15:20:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:58.543 [2024-11-19 15:20:48.868883] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:58.804 15:20:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:58.804 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 4f449536-f5b5-45e3-abab-fe44e561a816 '!=' 4f449536-f5b5-45e3-abab-fe44e561a816 ']' 00:14:58.804 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid5f 00:14:58.804 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:14:58.804 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:14:58.804 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:14:58.804 15:20:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:58.804 15:20:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:58.804 [2024-11-19 15:20:48.896721] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:14:58.804 15:20:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:58.804 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:58.804 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:58.804 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:58.804 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:58.804 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:58.804 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:58.804 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:58.804 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:58.804 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:58.804 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:58.804 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:58.804 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:58.804 15:20:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:58.804 15:20:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:58.804 15:20:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:58.804 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:58.804 "name": "raid_bdev1", 00:14:58.804 "uuid": "4f449536-f5b5-45e3-abab-fe44e561a816", 00:14:58.804 "strip_size_kb": 64, 00:14:58.804 "state": "online", 00:14:58.804 "raid_level": "raid5f", 00:14:58.804 "superblock": true, 00:14:58.804 "num_base_bdevs": 3, 00:14:58.804 "num_base_bdevs_discovered": 2, 00:14:58.804 "num_base_bdevs_operational": 2, 00:14:58.804 "base_bdevs_list": [ 00:14:58.804 { 00:14:58.804 "name": null, 00:14:58.804 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:58.804 "is_configured": false, 00:14:58.804 "data_offset": 0, 00:14:58.804 "data_size": 63488 00:14:58.804 }, 00:14:58.804 { 00:14:58.804 "name": "pt2", 00:14:58.804 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:58.804 "is_configured": true, 00:14:58.804 "data_offset": 2048, 00:14:58.804 "data_size": 63488 00:14:58.804 }, 00:14:58.804 { 00:14:58.804 "name": "pt3", 00:14:58.804 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:58.804 "is_configured": true, 00:14:58.804 "data_offset": 2048, 00:14:58.804 "data_size": 63488 00:14:58.804 } 00:14:58.804 ] 00:14:58.804 }' 00:14:58.804 15:20:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:58.804 15:20:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:59.065 [2024-11-19 15:20:49.288029] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:59.065 [2024-11-19 15:20:49.288100] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:59.065 [2024-11-19 15:20:49.288165] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:59.065 [2024-11-19 15:20:49.288243] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:59.065 [2024-11-19 15:20:49.288289] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:59.065 [2024-11-19 15:20:49.375866] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:14:59.065 [2024-11-19 15:20:49.375962] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:59.065 [2024-11-19 15:20:49.376030] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009980 00:14:59.065 [2024-11-19 15:20:49.376063] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:59.065 [2024-11-19 15:20:49.378227] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:59.065 [2024-11-19 15:20:49.378296] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:14:59.065 [2024-11-19 15:20:49.378399] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:14:59.065 [2024-11-19 15:20:49.378445] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:59.065 pt2 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 2 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:59.065 15:20:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:59.325 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:59.325 "name": "raid_bdev1", 00:14:59.325 "uuid": "4f449536-f5b5-45e3-abab-fe44e561a816", 00:14:59.325 "strip_size_kb": 64, 00:14:59.325 "state": "configuring", 00:14:59.325 "raid_level": "raid5f", 00:14:59.325 "superblock": true, 00:14:59.325 "num_base_bdevs": 3, 00:14:59.325 "num_base_bdevs_discovered": 1, 00:14:59.325 "num_base_bdevs_operational": 2, 00:14:59.325 "base_bdevs_list": [ 00:14:59.325 { 00:14:59.325 "name": null, 00:14:59.325 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:59.325 "is_configured": false, 00:14:59.325 "data_offset": 2048, 00:14:59.325 "data_size": 63488 00:14:59.325 }, 00:14:59.325 { 00:14:59.325 "name": "pt2", 00:14:59.325 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:59.325 "is_configured": true, 00:14:59.325 "data_offset": 2048, 00:14:59.325 "data_size": 63488 00:14:59.325 }, 00:14:59.325 { 00:14:59.325 "name": null, 00:14:59.325 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:59.325 "is_configured": false, 00:14:59.325 "data_offset": 2048, 00:14:59.325 "data_size": 63488 00:14:59.325 } 00:14:59.325 ] 00:14:59.325 }' 00:14:59.325 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:59.325 15:20:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:59.585 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:14:59.585 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:14:59.585 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@519 -- # i=2 00:14:59.585 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:14:59.585 15:20:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:59.585 15:20:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:59.585 [2024-11-19 15:20:49.807241] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:14:59.585 [2024-11-19 15:20:49.807336] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:59.585 [2024-11-19 15:20:49.807371] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:14:59.585 [2024-11-19 15:20:49.807398] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:59.585 [2024-11-19 15:20:49.807745] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:59.585 [2024-11-19 15:20:49.807809] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:14:59.585 [2024-11-19 15:20:49.807888] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:14:59.585 [2024-11-19 15:20:49.807932] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:14:59.586 [2024-11-19 15:20:49.808038] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:14:59.586 [2024-11-19 15:20:49.808074] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:14:59.586 [2024-11-19 15:20:49.808315] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:14:59.586 [2024-11-19 15:20:49.808787] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:14:59.586 [2024-11-19 15:20:49.808838] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:14:59.586 [2024-11-19 15:20:49.809107] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:59.586 pt3 00:14:59.586 15:20:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:59.586 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:59.586 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:59.586 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:59.586 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:59.586 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:59.586 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:59.586 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:59.586 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:59.586 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:59.586 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:59.586 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:59.586 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:59.586 15:20:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:59.586 15:20:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:59.586 15:20:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:59.586 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:59.586 "name": "raid_bdev1", 00:14:59.586 "uuid": "4f449536-f5b5-45e3-abab-fe44e561a816", 00:14:59.586 "strip_size_kb": 64, 00:14:59.586 "state": "online", 00:14:59.586 "raid_level": "raid5f", 00:14:59.586 "superblock": true, 00:14:59.586 "num_base_bdevs": 3, 00:14:59.586 "num_base_bdevs_discovered": 2, 00:14:59.586 "num_base_bdevs_operational": 2, 00:14:59.586 "base_bdevs_list": [ 00:14:59.586 { 00:14:59.586 "name": null, 00:14:59.586 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:59.586 "is_configured": false, 00:14:59.586 "data_offset": 2048, 00:14:59.586 "data_size": 63488 00:14:59.586 }, 00:14:59.586 { 00:14:59.586 "name": "pt2", 00:14:59.586 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:59.586 "is_configured": true, 00:14:59.586 "data_offset": 2048, 00:14:59.586 "data_size": 63488 00:14:59.586 }, 00:14:59.586 { 00:14:59.586 "name": "pt3", 00:14:59.586 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:59.586 "is_configured": true, 00:14:59.586 "data_offset": 2048, 00:14:59.586 "data_size": 63488 00:14:59.586 } 00:14:59.586 ] 00:14:59.586 }' 00:14:59.586 15:20:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:59.586 15:20:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:00.158 15:20:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:00.158 15:20:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:00.158 15:20:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:00.158 [2024-11-19 15:20:50.218538] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:00.158 [2024-11-19 15:20:50.218611] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:00.158 [2024-11-19 15:20:50.218683] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:00.158 [2024-11-19 15:20:50.218758] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:00.158 [2024-11-19 15:20:50.218808] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:15:00.158 15:20:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:00.158 15:20:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:00.158 15:20:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:00.158 15:20:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:15:00.158 15:20:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:00.158 15:20:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:00.158 15:20:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:15:00.158 15:20:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:15:00.158 15:20:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 3 -gt 2 ']' 00:15:00.158 15:20:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@534 -- # i=2 00:15:00.159 15:20:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt3 00:15:00.159 15:20:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:00.159 15:20:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:00.159 15:20:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:00.159 15:20:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:00.159 15:20:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:00.159 15:20:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:00.159 [2024-11-19 15:20:50.294412] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:00.159 [2024-11-19 15:20:50.294519] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:00.159 [2024-11-19 15:20:50.294549] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:15:00.159 [2024-11-19 15:20:50.294577] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:00.159 [2024-11-19 15:20:50.296698] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:00.159 [2024-11-19 15:20:50.296768] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:00.159 [2024-11-19 15:20:50.296844] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:15:00.159 [2024-11-19 15:20:50.296917] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:00.159 [2024-11-19 15:20:50.297054] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:15:00.159 [2024-11-19 15:20:50.297122] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:00.159 [2024-11-19 15:20:50.297164] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:15:00.159 [2024-11-19 15:20:50.297228] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:00.159 pt1 00:15:00.159 15:20:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:00.159 15:20:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 3 -gt 2 ']' 00:15:00.159 15:20:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 2 00:15:00.159 15:20:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:00.159 15:20:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:00.159 15:20:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:00.159 15:20:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:00.159 15:20:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:00.159 15:20:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:00.159 15:20:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:00.159 15:20:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:00.159 15:20:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:00.159 15:20:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:00.159 15:20:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:00.159 15:20:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:00.159 15:20:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:00.159 15:20:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:00.159 15:20:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:00.159 "name": "raid_bdev1", 00:15:00.159 "uuid": "4f449536-f5b5-45e3-abab-fe44e561a816", 00:15:00.159 "strip_size_kb": 64, 00:15:00.159 "state": "configuring", 00:15:00.159 "raid_level": "raid5f", 00:15:00.159 "superblock": true, 00:15:00.159 "num_base_bdevs": 3, 00:15:00.159 "num_base_bdevs_discovered": 1, 00:15:00.159 "num_base_bdevs_operational": 2, 00:15:00.159 "base_bdevs_list": [ 00:15:00.159 { 00:15:00.159 "name": null, 00:15:00.159 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:00.159 "is_configured": false, 00:15:00.159 "data_offset": 2048, 00:15:00.159 "data_size": 63488 00:15:00.159 }, 00:15:00.159 { 00:15:00.159 "name": "pt2", 00:15:00.159 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:00.159 "is_configured": true, 00:15:00.159 "data_offset": 2048, 00:15:00.159 "data_size": 63488 00:15:00.159 }, 00:15:00.159 { 00:15:00.159 "name": null, 00:15:00.159 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:00.159 "is_configured": false, 00:15:00.159 "data_offset": 2048, 00:15:00.159 "data_size": 63488 00:15:00.159 } 00:15:00.159 ] 00:15:00.159 }' 00:15:00.159 15:20:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:00.159 15:20:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:00.431 15:20:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:15:00.431 15:20:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:15:00.431 15:20:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:00.431 15:20:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:00.741 15:20:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:00.741 15:20:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:15:00.741 15:20:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:15:00.741 15:20:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:00.741 15:20:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:00.741 [2024-11-19 15:20:50.789621] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:15:00.741 [2024-11-19 15:20:50.789720] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:00.741 [2024-11-19 15:20:50.789751] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:15:00.741 [2024-11-19 15:20:50.789781] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:00.741 [2024-11-19 15:20:50.790150] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:00.741 [2024-11-19 15:20:50.790214] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:15:00.741 [2024-11-19 15:20:50.790291] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:15:00.741 [2024-11-19 15:20:50.790342] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:15:00.741 [2024-11-19 15:20:50.790446] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:15:00.741 [2024-11-19 15:20:50.790501] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:15:00.741 [2024-11-19 15:20:50.790740] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:15:00.741 [2024-11-19 15:20:50.791212] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:15:00.741 [2024-11-19 15:20:50.791261] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:15:00.741 [2024-11-19 15:20:50.791452] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:00.741 pt3 00:15:00.741 15:20:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:00.741 15:20:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:15:00.741 15:20:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:00.741 15:20:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:00.741 15:20:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:00.741 15:20:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:00.741 15:20:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:00.741 15:20:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:00.741 15:20:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:00.741 15:20:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:00.741 15:20:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:00.741 15:20:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:00.741 15:20:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:00.741 15:20:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:00.741 15:20:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:00.741 15:20:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:00.741 15:20:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:00.741 "name": "raid_bdev1", 00:15:00.741 "uuid": "4f449536-f5b5-45e3-abab-fe44e561a816", 00:15:00.741 "strip_size_kb": 64, 00:15:00.741 "state": "online", 00:15:00.741 "raid_level": "raid5f", 00:15:00.741 "superblock": true, 00:15:00.741 "num_base_bdevs": 3, 00:15:00.741 "num_base_bdevs_discovered": 2, 00:15:00.741 "num_base_bdevs_operational": 2, 00:15:00.741 "base_bdevs_list": [ 00:15:00.741 { 00:15:00.741 "name": null, 00:15:00.741 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:00.741 "is_configured": false, 00:15:00.741 "data_offset": 2048, 00:15:00.741 "data_size": 63488 00:15:00.741 }, 00:15:00.741 { 00:15:00.741 "name": "pt2", 00:15:00.741 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:00.741 "is_configured": true, 00:15:00.741 "data_offset": 2048, 00:15:00.741 "data_size": 63488 00:15:00.741 }, 00:15:00.741 { 00:15:00.741 "name": "pt3", 00:15:00.741 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:00.741 "is_configured": true, 00:15:00.741 "data_offset": 2048, 00:15:00.741 "data_size": 63488 00:15:00.741 } 00:15:00.741 ] 00:15:00.741 }' 00:15:00.741 15:20:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:00.741 15:20:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:01.002 15:20:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:15:01.002 15:20:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:15:01.002 15:20:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:01.002 15:20:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:01.002 15:20:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:01.002 15:20:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:15:01.002 15:20:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:01.002 15:20:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:15:01.002 15:20:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:01.002 15:20:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:01.002 [2024-11-19 15:20:51.265006] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:01.002 15:20:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:01.002 15:20:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' 4f449536-f5b5-45e3-abab-fe44e561a816 '!=' 4f449536-f5b5-45e3-abab-fe44e561a816 ']' 00:15:01.002 15:20:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 91671 00:15:01.002 15:20:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 91671 ']' 00:15:01.002 15:20:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@958 -- # kill -0 91671 00:15:01.002 15:20:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@959 -- # uname 00:15:01.002 15:20:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:15:01.002 15:20:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 91671 00:15:01.263 killing process with pid 91671 00:15:01.263 15:20:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:15:01.263 15:20:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:15:01.263 15:20:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 91671' 00:15:01.263 15:20:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@973 -- # kill 91671 00:15:01.263 [2024-11-19 15:20:51.343720] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:01.263 [2024-11-19 15:20:51.343799] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:01.263 [2024-11-19 15:20:51.343851] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:01.263 [2024-11-19 15:20:51.343860] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:15:01.263 15:20:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@978 -- # wait 91671 00:15:01.263 [2024-11-19 15:20:51.376817] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:01.263 ************************************ 00:15:01.263 END TEST raid5f_superblock_test 00:15:01.263 ************************************ 00:15:01.263 15:20:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:15:01.263 00:15:01.263 real 0m6.265s 00:15:01.263 user 0m10.465s 00:15:01.263 sys 0m1.411s 00:15:01.263 15:20:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:15:01.263 15:20:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:01.523 15:20:51 bdev_raid -- bdev/bdev_raid.sh@989 -- # '[' true = true ']' 00:15:01.523 15:20:51 bdev_raid -- bdev/bdev_raid.sh@990 -- # run_test raid5f_rebuild_test raid_rebuild_test raid5f 3 false false true 00:15:01.523 15:20:51 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:15:01.523 15:20:51 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:15:01.523 15:20:51 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:01.523 ************************************ 00:15:01.523 START TEST raid5f_rebuild_test 00:15:01.523 ************************************ 00:15:01.523 15:20:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid5f 3 false false true 00:15:01.523 15:20:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:15:01.523 15:20:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=3 00:15:01.523 15:20:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:15:01.523 15:20:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:15:01.523 15:20:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:15:01.523 15:20:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:15:01.523 15:20:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:01.523 15:20:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:15:01.523 15:20:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:01.523 15:20:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:01.523 15:20:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:15:01.523 15:20:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:01.523 15:20:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:01.523 15:20:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:15:01.523 15:20:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:01.523 15:20:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:01.523 15:20:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:15:01.523 15:20:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:15:01.523 15:20:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:15:01.523 15:20:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:15:01.523 15:20:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:15:01.523 15:20:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:15:01.523 15:20:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:15:01.523 15:20:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:15:01.523 15:20:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:15:01.523 15:20:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:15:01.523 15:20:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:15:01.524 15:20:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:15:01.524 15:20:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=92099 00:15:01.524 15:20:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:15:01.524 15:20:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 92099 00:15:01.524 15:20:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@835 -- # '[' -z 92099 ']' 00:15:01.524 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:01.524 15:20:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:01.524 15:20:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:15:01.524 15:20:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:01.524 15:20:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:15:01.524 15:20:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:01.524 [2024-11-19 15:20:51.767044] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:15:01.524 I/O size of 3145728 is greater than zero copy threshold (65536). 00:15:01.524 Zero copy mechanism will not be used. 00:15:01.524 [2024-11-19 15:20:51.767272] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid92099 ] 00:15:01.783 [2024-11-19 15:20:51.924499] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:01.783 [2024-11-19 15:20:51.950023] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:01.783 [2024-11-19 15:20:51.993002] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:01.783 [2024-11-19 15:20:51.993134] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:02.353 15:20:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:15:02.353 15:20:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@868 -- # return 0 00:15:02.353 15:20:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:02.353 15:20:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:15:02.353 15:20:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:02.353 15:20:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:02.353 BaseBdev1_malloc 00:15:02.353 15:20:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:02.353 15:20:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:15:02.353 15:20:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:02.353 15:20:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:02.353 [2024-11-19 15:20:52.596093] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:15:02.353 [2024-11-19 15:20:52.596156] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:02.353 [2024-11-19 15:20:52.596183] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:15:02.353 [2024-11-19 15:20:52.596195] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:02.353 [2024-11-19 15:20:52.598328] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:02.353 [2024-11-19 15:20:52.598439] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:15:02.353 BaseBdev1 00:15:02.353 15:20:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:02.353 15:20:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:02.353 15:20:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:15:02.353 15:20:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:02.353 15:20:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:02.353 BaseBdev2_malloc 00:15:02.353 15:20:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:02.353 15:20:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:15:02.353 15:20:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:02.353 15:20:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:02.353 [2024-11-19 15:20:52.624732] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:15:02.353 [2024-11-19 15:20:52.624776] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:02.353 [2024-11-19 15:20:52.624796] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:15:02.353 [2024-11-19 15:20:52.624804] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:02.353 [2024-11-19 15:20:52.626924] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:02.353 [2024-11-19 15:20:52.626970] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:15:02.353 BaseBdev2 00:15:02.353 15:20:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:02.353 15:20:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:02.353 15:20:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:15:02.353 15:20:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:02.353 15:20:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:02.353 BaseBdev3_malloc 00:15:02.353 15:20:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:02.353 15:20:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:15:02.353 15:20:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:02.353 15:20:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:02.353 [2024-11-19 15:20:52.653233] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:15:02.353 [2024-11-19 15:20:52.653282] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:02.353 [2024-11-19 15:20:52.653302] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:15:02.353 [2024-11-19 15:20:52.653311] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:02.353 [2024-11-19 15:20:52.655350] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:02.353 [2024-11-19 15:20:52.655383] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:15:02.353 BaseBdev3 00:15:02.353 15:20:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:02.353 15:20:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:15:02.353 15:20:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:02.353 15:20:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:02.613 spare_malloc 00:15:02.613 15:20:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:02.613 15:20:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:15:02.613 15:20:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:02.613 15:20:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:02.613 spare_delay 00:15:02.613 15:20:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:02.613 15:20:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:02.613 15:20:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:02.613 15:20:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:02.613 [2024-11-19 15:20:52.712262] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:02.613 [2024-11-19 15:20:52.712322] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:02.613 [2024-11-19 15:20:52.712354] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:15:02.613 [2024-11-19 15:20:52.712364] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:02.613 [2024-11-19 15:20:52.714764] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:02.614 [2024-11-19 15:20:52.714803] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:02.614 spare 00:15:02.614 15:20:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:02.614 15:20:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 00:15:02.614 15:20:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:02.614 15:20:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:02.614 [2024-11-19 15:20:52.724291] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:02.614 [2024-11-19 15:20:52.726109] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:02.614 [2024-11-19 15:20:52.726168] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:15:02.614 [2024-11-19 15:20:52.726262] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:15:02.614 [2024-11-19 15:20:52.726273] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:15:02.614 [2024-11-19 15:20:52.726532] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:15:02.614 [2024-11-19 15:20:52.726942] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:15:02.614 [2024-11-19 15:20:52.726952] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:15:02.614 [2024-11-19 15:20:52.727086] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:02.614 15:20:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:02.614 15:20:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:02.614 15:20:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:02.614 15:20:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:02.614 15:20:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:02.614 15:20:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:02.614 15:20:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:02.614 15:20:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:02.614 15:20:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:02.614 15:20:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:02.614 15:20:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:02.614 15:20:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:02.614 15:20:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:02.614 15:20:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:02.614 15:20:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:02.614 15:20:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:02.614 15:20:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:02.614 "name": "raid_bdev1", 00:15:02.614 "uuid": "d41199a8-b2e7-484a-a502-e4fec162c91e", 00:15:02.614 "strip_size_kb": 64, 00:15:02.614 "state": "online", 00:15:02.614 "raid_level": "raid5f", 00:15:02.614 "superblock": false, 00:15:02.614 "num_base_bdevs": 3, 00:15:02.614 "num_base_bdevs_discovered": 3, 00:15:02.614 "num_base_bdevs_operational": 3, 00:15:02.614 "base_bdevs_list": [ 00:15:02.614 { 00:15:02.614 "name": "BaseBdev1", 00:15:02.614 "uuid": "b696053a-a6af-58c3-8183-64cbd02e9b3a", 00:15:02.614 "is_configured": true, 00:15:02.614 "data_offset": 0, 00:15:02.614 "data_size": 65536 00:15:02.614 }, 00:15:02.614 { 00:15:02.614 "name": "BaseBdev2", 00:15:02.614 "uuid": "0ae5b1b0-1bb6-53af-83e2-1c19cd05b07b", 00:15:02.614 "is_configured": true, 00:15:02.614 "data_offset": 0, 00:15:02.614 "data_size": 65536 00:15:02.614 }, 00:15:02.614 { 00:15:02.614 "name": "BaseBdev3", 00:15:02.614 "uuid": "a64b1081-9e6a-5aba-adb8-dd4b0b2a2092", 00:15:02.614 "is_configured": true, 00:15:02.614 "data_offset": 0, 00:15:02.614 "data_size": 65536 00:15:02.614 } 00:15:02.614 ] 00:15:02.614 }' 00:15:02.614 15:20:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:02.614 15:20:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:02.874 15:20:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:02.874 15:20:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:02.874 15:20:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:15:02.874 15:20:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:02.874 [2024-11-19 15:20:53.208056] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:03.133 15:20:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:03.133 15:20:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=131072 00:15:03.133 15:20:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:03.133 15:20:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:15:03.133 15:20:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:03.133 15:20:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:03.133 15:20:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:03.133 15:20:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:15:03.133 15:20:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:15:03.133 15:20:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:15:03.133 15:20:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:15:03.133 15:20:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:15:03.133 15:20:53 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:15:03.133 15:20:53 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:15:03.133 15:20:53 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:15:03.133 15:20:53 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:15:03.133 15:20:53 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:15:03.133 15:20:53 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:15:03.133 15:20:53 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:15:03.133 15:20:53 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:15:03.133 15:20:53 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:15:03.393 [2024-11-19 15:20:53.483573] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:15:03.393 /dev/nbd0 00:15:03.393 15:20:53 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:15:03.393 15:20:53 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:15:03.393 15:20:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:15:03.393 15:20:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:15:03.393 15:20:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:15:03.393 15:20:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:15:03.393 15:20:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:15:03.393 15:20:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@877 -- # break 00:15:03.393 15:20:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:15:03.393 15:20:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:15:03.393 15:20:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:03.393 1+0 records in 00:15:03.393 1+0 records out 00:15:03.393 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000456729 s, 9.0 MB/s 00:15:03.393 15:20:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:03.393 15:20:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:15:03.393 15:20:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:03.393 15:20:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:15:03.393 15:20:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:15:03.393 15:20:53 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:03.393 15:20:53 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:15:03.393 15:20:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:15:03.393 15:20:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@630 -- # write_unit_size=256 00:15:03.393 15:20:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@631 -- # echo 128 00:15:03.393 15:20:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=131072 count=512 oflag=direct 00:15:03.654 512+0 records in 00:15:03.654 512+0 records out 00:15:03.654 67108864 bytes (67 MB, 64 MiB) copied, 0.300758 s, 223 MB/s 00:15:03.654 15:20:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:15:03.654 15:20:53 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:15:03.654 15:20:53 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:15:03.654 15:20:53 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:15:03.654 15:20:53 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:15:03.654 15:20:53 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:03.654 15:20:53 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:15:03.914 15:20:54 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:15:03.914 [2024-11-19 15:20:54.086519] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:03.914 15:20:54 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:15:03.914 15:20:54 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:15:03.914 15:20:54 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:03.914 15:20:54 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:03.914 15:20:54 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:15:03.914 15:20:54 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:15:03.914 15:20:54 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:15:03.914 15:20:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:15:03.914 15:20:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:03.914 15:20:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:03.914 [2024-11-19 15:20:54.102491] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:03.914 15:20:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:03.914 15:20:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:15:03.914 15:20:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:03.914 15:20:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:03.914 15:20:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:03.914 15:20:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:03.914 15:20:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:03.914 15:20:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:03.914 15:20:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:03.914 15:20:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:03.914 15:20:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:03.914 15:20:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:03.914 15:20:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:03.914 15:20:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:03.914 15:20:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:03.914 15:20:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:03.914 15:20:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:03.914 "name": "raid_bdev1", 00:15:03.914 "uuid": "d41199a8-b2e7-484a-a502-e4fec162c91e", 00:15:03.914 "strip_size_kb": 64, 00:15:03.914 "state": "online", 00:15:03.914 "raid_level": "raid5f", 00:15:03.914 "superblock": false, 00:15:03.914 "num_base_bdevs": 3, 00:15:03.914 "num_base_bdevs_discovered": 2, 00:15:03.914 "num_base_bdevs_operational": 2, 00:15:03.914 "base_bdevs_list": [ 00:15:03.914 { 00:15:03.914 "name": null, 00:15:03.914 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:03.914 "is_configured": false, 00:15:03.914 "data_offset": 0, 00:15:03.914 "data_size": 65536 00:15:03.914 }, 00:15:03.914 { 00:15:03.914 "name": "BaseBdev2", 00:15:03.914 "uuid": "0ae5b1b0-1bb6-53af-83e2-1c19cd05b07b", 00:15:03.914 "is_configured": true, 00:15:03.914 "data_offset": 0, 00:15:03.914 "data_size": 65536 00:15:03.915 }, 00:15:03.915 { 00:15:03.915 "name": "BaseBdev3", 00:15:03.915 "uuid": "a64b1081-9e6a-5aba-adb8-dd4b0b2a2092", 00:15:03.915 "is_configured": true, 00:15:03.915 "data_offset": 0, 00:15:03.915 "data_size": 65536 00:15:03.915 } 00:15:03.915 ] 00:15:03.915 }' 00:15:03.915 15:20:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:03.915 15:20:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:04.484 15:20:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:04.484 15:20:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:04.485 15:20:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:04.485 [2024-11-19 15:20:54.577672] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:04.485 [2024-11-19 15:20:54.582329] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000027cd0 00:15:04.485 15:20:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:04.485 15:20:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:15:04.485 [2024-11-19 15:20:54.584534] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:05.424 15:20:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:05.424 15:20:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:05.424 15:20:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:05.424 15:20:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:05.424 15:20:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:05.424 15:20:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:05.424 15:20:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:05.424 15:20:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:05.424 15:20:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:05.424 15:20:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:05.424 15:20:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:05.424 "name": "raid_bdev1", 00:15:05.424 "uuid": "d41199a8-b2e7-484a-a502-e4fec162c91e", 00:15:05.424 "strip_size_kb": 64, 00:15:05.424 "state": "online", 00:15:05.424 "raid_level": "raid5f", 00:15:05.424 "superblock": false, 00:15:05.424 "num_base_bdevs": 3, 00:15:05.424 "num_base_bdevs_discovered": 3, 00:15:05.424 "num_base_bdevs_operational": 3, 00:15:05.424 "process": { 00:15:05.425 "type": "rebuild", 00:15:05.425 "target": "spare", 00:15:05.425 "progress": { 00:15:05.425 "blocks": 20480, 00:15:05.425 "percent": 15 00:15:05.425 } 00:15:05.425 }, 00:15:05.425 "base_bdevs_list": [ 00:15:05.425 { 00:15:05.425 "name": "spare", 00:15:05.425 "uuid": "ae650596-5d7a-5db8-9c0b-c2e81660fafd", 00:15:05.425 "is_configured": true, 00:15:05.425 "data_offset": 0, 00:15:05.425 "data_size": 65536 00:15:05.425 }, 00:15:05.425 { 00:15:05.425 "name": "BaseBdev2", 00:15:05.425 "uuid": "0ae5b1b0-1bb6-53af-83e2-1c19cd05b07b", 00:15:05.425 "is_configured": true, 00:15:05.425 "data_offset": 0, 00:15:05.425 "data_size": 65536 00:15:05.425 }, 00:15:05.425 { 00:15:05.425 "name": "BaseBdev3", 00:15:05.425 "uuid": "a64b1081-9e6a-5aba-adb8-dd4b0b2a2092", 00:15:05.425 "is_configured": true, 00:15:05.425 "data_offset": 0, 00:15:05.425 "data_size": 65536 00:15:05.425 } 00:15:05.425 ] 00:15:05.425 }' 00:15:05.425 15:20:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:05.425 15:20:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:05.425 15:20:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:05.425 15:20:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:05.425 15:20:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:15:05.425 15:20:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:05.425 15:20:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:05.425 [2024-11-19 15:20:55.752441] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:05.685 [2024-11-19 15:20:55.791118] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:05.685 [2024-11-19 15:20:55.791173] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:05.685 [2024-11-19 15:20:55.791187] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:05.685 [2024-11-19 15:20:55.791199] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:05.685 15:20:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:05.685 15:20:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:15:05.685 15:20:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:05.685 15:20:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:05.685 15:20:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:05.685 15:20:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:05.685 15:20:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:05.685 15:20:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:05.685 15:20:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:05.685 15:20:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:05.685 15:20:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:05.685 15:20:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:05.685 15:20:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:05.685 15:20:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:05.685 15:20:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:05.685 15:20:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:05.685 15:20:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:05.685 "name": "raid_bdev1", 00:15:05.685 "uuid": "d41199a8-b2e7-484a-a502-e4fec162c91e", 00:15:05.685 "strip_size_kb": 64, 00:15:05.685 "state": "online", 00:15:05.685 "raid_level": "raid5f", 00:15:05.685 "superblock": false, 00:15:05.685 "num_base_bdevs": 3, 00:15:05.685 "num_base_bdevs_discovered": 2, 00:15:05.685 "num_base_bdevs_operational": 2, 00:15:05.685 "base_bdevs_list": [ 00:15:05.685 { 00:15:05.685 "name": null, 00:15:05.685 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:05.685 "is_configured": false, 00:15:05.685 "data_offset": 0, 00:15:05.685 "data_size": 65536 00:15:05.685 }, 00:15:05.685 { 00:15:05.685 "name": "BaseBdev2", 00:15:05.685 "uuid": "0ae5b1b0-1bb6-53af-83e2-1c19cd05b07b", 00:15:05.685 "is_configured": true, 00:15:05.685 "data_offset": 0, 00:15:05.685 "data_size": 65536 00:15:05.685 }, 00:15:05.685 { 00:15:05.685 "name": "BaseBdev3", 00:15:05.685 "uuid": "a64b1081-9e6a-5aba-adb8-dd4b0b2a2092", 00:15:05.685 "is_configured": true, 00:15:05.685 "data_offset": 0, 00:15:05.685 "data_size": 65536 00:15:05.685 } 00:15:05.685 ] 00:15:05.685 }' 00:15:05.685 15:20:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:05.685 15:20:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:05.945 15:20:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:05.945 15:20:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:05.945 15:20:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:05.945 15:20:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:05.945 15:20:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:05.945 15:20:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:05.945 15:20:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:05.945 15:20:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:05.945 15:20:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:06.205 15:20:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:06.205 15:20:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:06.205 "name": "raid_bdev1", 00:15:06.205 "uuid": "d41199a8-b2e7-484a-a502-e4fec162c91e", 00:15:06.205 "strip_size_kb": 64, 00:15:06.205 "state": "online", 00:15:06.205 "raid_level": "raid5f", 00:15:06.205 "superblock": false, 00:15:06.205 "num_base_bdevs": 3, 00:15:06.205 "num_base_bdevs_discovered": 2, 00:15:06.205 "num_base_bdevs_operational": 2, 00:15:06.205 "base_bdevs_list": [ 00:15:06.205 { 00:15:06.205 "name": null, 00:15:06.205 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:06.205 "is_configured": false, 00:15:06.205 "data_offset": 0, 00:15:06.205 "data_size": 65536 00:15:06.205 }, 00:15:06.205 { 00:15:06.205 "name": "BaseBdev2", 00:15:06.205 "uuid": "0ae5b1b0-1bb6-53af-83e2-1c19cd05b07b", 00:15:06.205 "is_configured": true, 00:15:06.205 "data_offset": 0, 00:15:06.205 "data_size": 65536 00:15:06.205 }, 00:15:06.205 { 00:15:06.205 "name": "BaseBdev3", 00:15:06.205 "uuid": "a64b1081-9e6a-5aba-adb8-dd4b0b2a2092", 00:15:06.205 "is_configured": true, 00:15:06.205 "data_offset": 0, 00:15:06.205 "data_size": 65536 00:15:06.205 } 00:15:06.205 ] 00:15:06.205 }' 00:15:06.205 15:20:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:06.205 15:20:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:06.205 15:20:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:06.205 15:20:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:06.205 15:20:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:06.205 15:20:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:06.205 15:20:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:06.205 [2024-11-19 15:20:56.407883] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:06.205 [2024-11-19 15:20:56.411547] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000027da0 00:15:06.205 [2024-11-19 15:20:56.413608] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:06.205 15:20:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:06.205 15:20:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:15:07.145 15:20:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:07.145 15:20:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:07.145 15:20:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:07.145 15:20:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:07.145 15:20:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:07.145 15:20:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:07.145 15:20:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:07.145 15:20:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:07.145 15:20:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:07.145 15:20:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:07.145 15:20:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:07.145 "name": "raid_bdev1", 00:15:07.145 "uuid": "d41199a8-b2e7-484a-a502-e4fec162c91e", 00:15:07.145 "strip_size_kb": 64, 00:15:07.145 "state": "online", 00:15:07.145 "raid_level": "raid5f", 00:15:07.145 "superblock": false, 00:15:07.145 "num_base_bdevs": 3, 00:15:07.145 "num_base_bdevs_discovered": 3, 00:15:07.145 "num_base_bdevs_operational": 3, 00:15:07.145 "process": { 00:15:07.145 "type": "rebuild", 00:15:07.145 "target": "spare", 00:15:07.145 "progress": { 00:15:07.145 "blocks": 20480, 00:15:07.145 "percent": 15 00:15:07.145 } 00:15:07.145 }, 00:15:07.145 "base_bdevs_list": [ 00:15:07.145 { 00:15:07.145 "name": "spare", 00:15:07.145 "uuid": "ae650596-5d7a-5db8-9c0b-c2e81660fafd", 00:15:07.145 "is_configured": true, 00:15:07.145 "data_offset": 0, 00:15:07.145 "data_size": 65536 00:15:07.145 }, 00:15:07.145 { 00:15:07.145 "name": "BaseBdev2", 00:15:07.145 "uuid": "0ae5b1b0-1bb6-53af-83e2-1c19cd05b07b", 00:15:07.145 "is_configured": true, 00:15:07.145 "data_offset": 0, 00:15:07.145 "data_size": 65536 00:15:07.145 }, 00:15:07.145 { 00:15:07.145 "name": "BaseBdev3", 00:15:07.145 "uuid": "a64b1081-9e6a-5aba-adb8-dd4b0b2a2092", 00:15:07.145 "is_configured": true, 00:15:07.145 "data_offset": 0, 00:15:07.145 "data_size": 65536 00:15:07.145 } 00:15:07.145 ] 00:15:07.145 }' 00:15:07.145 15:20:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:07.406 15:20:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:07.406 15:20:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:07.406 15:20:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:07.406 15:20:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:15:07.406 15:20:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=3 00:15:07.406 15:20:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:15:07.406 15:20:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=455 00:15:07.406 15:20:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:07.406 15:20:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:07.406 15:20:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:07.406 15:20:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:07.406 15:20:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:07.406 15:20:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:07.406 15:20:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:07.406 15:20:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:07.406 15:20:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:07.406 15:20:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:07.406 15:20:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:07.406 15:20:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:07.406 "name": "raid_bdev1", 00:15:07.406 "uuid": "d41199a8-b2e7-484a-a502-e4fec162c91e", 00:15:07.406 "strip_size_kb": 64, 00:15:07.406 "state": "online", 00:15:07.406 "raid_level": "raid5f", 00:15:07.406 "superblock": false, 00:15:07.406 "num_base_bdevs": 3, 00:15:07.406 "num_base_bdevs_discovered": 3, 00:15:07.406 "num_base_bdevs_operational": 3, 00:15:07.406 "process": { 00:15:07.406 "type": "rebuild", 00:15:07.406 "target": "spare", 00:15:07.406 "progress": { 00:15:07.406 "blocks": 22528, 00:15:07.406 "percent": 17 00:15:07.406 } 00:15:07.406 }, 00:15:07.406 "base_bdevs_list": [ 00:15:07.406 { 00:15:07.406 "name": "spare", 00:15:07.406 "uuid": "ae650596-5d7a-5db8-9c0b-c2e81660fafd", 00:15:07.406 "is_configured": true, 00:15:07.406 "data_offset": 0, 00:15:07.406 "data_size": 65536 00:15:07.406 }, 00:15:07.406 { 00:15:07.406 "name": "BaseBdev2", 00:15:07.406 "uuid": "0ae5b1b0-1bb6-53af-83e2-1c19cd05b07b", 00:15:07.406 "is_configured": true, 00:15:07.406 "data_offset": 0, 00:15:07.406 "data_size": 65536 00:15:07.406 }, 00:15:07.406 { 00:15:07.406 "name": "BaseBdev3", 00:15:07.406 "uuid": "a64b1081-9e6a-5aba-adb8-dd4b0b2a2092", 00:15:07.406 "is_configured": true, 00:15:07.406 "data_offset": 0, 00:15:07.406 "data_size": 65536 00:15:07.406 } 00:15:07.406 ] 00:15:07.406 }' 00:15:07.406 15:20:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:07.406 15:20:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:07.406 15:20:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:07.406 15:20:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:07.406 15:20:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:08.789 15:20:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:08.789 15:20:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:08.789 15:20:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:08.789 15:20:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:08.789 15:20:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:08.789 15:20:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:08.789 15:20:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:08.789 15:20:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:08.789 15:20:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:08.789 15:20:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:08.789 15:20:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:08.789 15:20:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:08.789 "name": "raid_bdev1", 00:15:08.789 "uuid": "d41199a8-b2e7-484a-a502-e4fec162c91e", 00:15:08.789 "strip_size_kb": 64, 00:15:08.789 "state": "online", 00:15:08.789 "raid_level": "raid5f", 00:15:08.789 "superblock": false, 00:15:08.789 "num_base_bdevs": 3, 00:15:08.789 "num_base_bdevs_discovered": 3, 00:15:08.789 "num_base_bdevs_operational": 3, 00:15:08.789 "process": { 00:15:08.789 "type": "rebuild", 00:15:08.789 "target": "spare", 00:15:08.789 "progress": { 00:15:08.789 "blocks": 47104, 00:15:08.789 "percent": 35 00:15:08.789 } 00:15:08.789 }, 00:15:08.789 "base_bdevs_list": [ 00:15:08.789 { 00:15:08.789 "name": "spare", 00:15:08.789 "uuid": "ae650596-5d7a-5db8-9c0b-c2e81660fafd", 00:15:08.789 "is_configured": true, 00:15:08.789 "data_offset": 0, 00:15:08.789 "data_size": 65536 00:15:08.789 }, 00:15:08.789 { 00:15:08.789 "name": "BaseBdev2", 00:15:08.789 "uuid": "0ae5b1b0-1bb6-53af-83e2-1c19cd05b07b", 00:15:08.789 "is_configured": true, 00:15:08.789 "data_offset": 0, 00:15:08.789 "data_size": 65536 00:15:08.789 }, 00:15:08.789 { 00:15:08.789 "name": "BaseBdev3", 00:15:08.789 "uuid": "a64b1081-9e6a-5aba-adb8-dd4b0b2a2092", 00:15:08.789 "is_configured": true, 00:15:08.789 "data_offset": 0, 00:15:08.789 "data_size": 65536 00:15:08.789 } 00:15:08.789 ] 00:15:08.789 }' 00:15:08.789 15:20:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:08.789 15:20:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:08.789 15:20:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:08.789 15:20:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:08.789 15:20:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:09.755 15:20:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:09.755 15:20:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:09.755 15:20:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:09.755 15:20:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:09.755 15:20:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:09.755 15:20:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:09.755 15:20:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:09.755 15:20:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:09.755 15:20:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:09.755 15:20:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:09.755 15:20:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:09.755 15:20:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:09.755 "name": "raid_bdev1", 00:15:09.755 "uuid": "d41199a8-b2e7-484a-a502-e4fec162c91e", 00:15:09.755 "strip_size_kb": 64, 00:15:09.755 "state": "online", 00:15:09.755 "raid_level": "raid5f", 00:15:09.755 "superblock": false, 00:15:09.755 "num_base_bdevs": 3, 00:15:09.755 "num_base_bdevs_discovered": 3, 00:15:09.755 "num_base_bdevs_operational": 3, 00:15:09.755 "process": { 00:15:09.755 "type": "rebuild", 00:15:09.755 "target": "spare", 00:15:09.755 "progress": { 00:15:09.755 "blocks": 69632, 00:15:09.755 "percent": 53 00:15:09.755 } 00:15:09.755 }, 00:15:09.755 "base_bdevs_list": [ 00:15:09.755 { 00:15:09.755 "name": "spare", 00:15:09.755 "uuid": "ae650596-5d7a-5db8-9c0b-c2e81660fafd", 00:15:09.755 "is_configured": true, 00:15:09.755 "data_offset": 0, 00:15:09.755 "data_size": 65536 00:15:09.755 }, 00:15:09.755 { 00:15:09.755 "name": "BaseBdev2", 00:15:09.755 "uuid": "0ae5b1b0-1bb6-53af-83e2-1c19cd05b07b", 00:15:09.755 "is_configured": true, 00:15:09.755 "data_offset": 0, 00:15:09.755 "data_size": 65536 00:15:09.755 }, 00:15:09.755 { 00:15:09.755 "name": "BaseBdev3", 00:15:09.755 "uuid": "a64b1081-9e6a-5aba-adb8-dd4b0b2a2092", 00:15:09.755 "is_configured": true, 00:15:09.755 "data_offset": 0, 00:15:09.755 "data_size": 65536 00:15:09.755 } 00:15:09.755 ] 00:15:09.755 }' 00:15:09.755 15:20:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:09.755 15:20:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:09.755 15:20:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:09.755 15:21:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:09.755 15:21:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:11.138 15:21:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:11.138 15:21:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:11.138 15:21:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:11.138 15:21:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:11.138 15:21:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:11.138 15:21:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:11.138 15:21:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:11.138 15:21:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:11.138 15:21:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:11.138 15:21:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:11.138 15:21:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:11.138 15:21:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:11.138 "name": "raid_bdev1", 00:15:11.138 "uuid": "d41199a8-b2e7-484a-a502-e4fec162c91e", 00:15:11.138 "strip_size_kb": 64, 00:15:11.138 "state": "online", 00:15:11.138 "raid_level": "raid5f", 00:15:11.138 "superblock": false, 00:15:11.138 "num_base_bdevs": 3, 00:15:11.138 "num_base_bdevs_discovered": 3, 00:15:11.138 "num_base_bdevs_operational": 3, 00:15:11.138 "process": { 00:15:11.138 "type": "rebuild", 00:15:11.138 "target": "spare", 00:15:11.138 "progress": { 00:15:11.138 "blocks": 94208, 00:15:11.138 "percent": 71 00:15:11.138 } 00:15:11.138 }, 00:15:11.138 "base_bdevs_list": [ 00:15:11.138 { 00:15:11.138 "name": "spare", 00:15:11.138 "uuid": "ae650596-5d7a-5db8-9c0b-c2e81660fafd", 00:15:11.138 "is_configured": true, 00:15:11.138 "data_offset": 0, 00:15:11.138 "data_size": 65536 00:15:11.138 }, 00:15:11.138 { 00:15:11.138 "name": "BaseBdev2", 00:15:11.138 "uuid": "0ae5b1b0-1bb6-53af-83e2-1c19cd05b07b", 00:15:11.138 "is_configured": true, 00:15:11.138 "data_offset": 0, 00:15:11.138 "data_size": 65536 00:15:11.138 }, 00:15:11.138 { 00:15:11.138 "name": "BaseBdev3", 00:15:11.138 "uuid": "a64b1081-9e6a-5aba-adb8-dd4b0b2a2092", 00:15:11.138 "is_configured": true, 00:15:11.138 "data_offset": 0, 00:15:11.138 "data_size": 65536 00:15:11.138 } 00:15:11.138 ] 00:15:11.138 }' 00:15:11.138 15:21:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:11.139 15:21:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:11.139 15:21:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:11.139 15:21:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:11.139 15:21:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:12.079 15:21:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:12.079 15:21:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:12.079 15:21:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:12.079 15:21:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:12.079 15:21:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:12.079 15:21:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:12.079 15:21:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:12.079 15:21:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:12.079 15:21:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:12.079 15:21:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:12.079 15:21:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:12.079 15:21:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:12.079 "name": "raid_bdev1", 00:15:12.079 "uuid": "d41199a8-b2e7-484a-a502-e4fec162c91e", 00:15:12.079 "strip_size_kb": 64, 00:15:12.079 "state": "online", 00:15:12.079 "raid_level": "raid5f", 00:15:12.079 "superblock": false, 00:15:12.079 "num_base_bdevs": 3, 00:15:12.079 "num_base_bdevs_discovered": 3, 00:15:12.079 "num_base_bdevs_operational": 3, 00:15:12.079 "process": { 00:15:12.079 "type": "rebuild", 00:15:12.079 "target": "spare", 00:15:12.079 "progress": { 00:15:12.079 "blocks": 116736, 00:15:12.079 "percent": 89 00:15:12.079 } 00:15:12.079 }, 00:15:12.079 "base_bdevs_list": [ 00:15:12.079 { 00:15:12.079 "name": "spare", 00:15:12.079 "uuid": "ae650596-5d7a-5db8-9c0b-c2e81660fafd", 00:15:12.079 "is_configured": true, 00:15:12.079 "data_offset": 0, 00:15:12.079 "data_size": 65536 00:15:12.079 }, 00:15:12.079 { 00:15:12.079 "name": "BaseBdev2", 00:15:12.079 "uuid": "0ae5b1b0-1bb6-53af-83e2-1c19cd05b07b", 00:15:12.079 "is_configured": true, 00:15:12.079 "data_offset": 0, 00:15:12.079 "data_size": 65536 00:15:12.079 }, 00:15:12.079 { 00:15:12.079 "name": "BaseBdev3", 00:15:12.079 "uuid": "a64b1081-9e6a-5aba-adb8-dd4b0b2a2092", 00:15:12.079 "is_configured": true, 00:15:12.079 "data_offset": 0, 00:15:12.079 "data_size": 65536 00:15:12.079 } 00:15:12.079 ] 00:15:12.079 }' 00:15:12.079 15:21:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:12.079 15:21:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:12.079 15:21:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:12.079 15:21:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:12.079 15:21:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:12.649 [2024-11-19 15:21:02.845925] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:15:12.649 [2024-11-19 15:21:02.846000] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:15:12.649 [2024-11-19 15:21:02.846040] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:13.219 15:21:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:13.219 15:21:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:13.219 15:21:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:13.219 15:21:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:13.219 15:21:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:13.219 15:21:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:13.219 15:21:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:13.219 15:21:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:13.219 15:21:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:13.219 15:21:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:13.219 15:21:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:13.219 15:21:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:13.219 "name": "raid_bdev1", 00:15:13.219 "uuid": "d41199a8-b2e7-484a-a502-e4fec162c91e", 00:15:13.219 "strip_size_kb": 64, 00:15:13.219 "state": "online", 00:15:13.219 "raid_level": "raid5f", 00:15:13.219 "superblock": false, 00:15:13.219 "num_base_bdevs": 3, 00:15:13.219 "num_base_bdevs_discovered": 3, 00:15:13.219 "num_base_bdevs_operational": 3, 00:15:13.219 "base_bdevs_list": [ 00:15:13.219 { 00:15:13.219 "name": "spare", 00:15:13.219 "uuid": "ae650596-5d7a-5db8-9c0b-c2e81660fafd", 00:15:13.219 "is_configured": true, 00:15:13.219 "data_offset": 0, 00:15:13.219 "data_size": 65536 00:15:13.219 }, 00:15:13.219 { 00:15:13.219 "name": "BaseBdev2", 00:15:13.219 "uuid": "0ae5b1b0-1bb6-53af-83e2-1c19cd05b07b", 00:15:13.219 "is_configured": true, 00:15:13.219 "data_offset": 0, 00:15:13.219 "data_size": 65536 00:15:13.219 }, 00:15:13.219 { 00:15:13.219 "name": "BaseBdev3", 00:15:13.219 "uuid": "a64b1081-9e6a-5aba-adb8-dd4b0b2a2092", 00:15:13.219 "is_configured": true, 00:15:13.219 "data_offset": 0, 00:15:13.219 "data_size": 65536 00:15:13.219 } 00:15:13.219 ] 00:15:13.219 }' 00:15:13.219 15:21:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:13.219 15:21:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:15:13.219 15:21:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:13.219 15:21:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:15:13.219 15:21:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:15:13.219 15:21:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:13.219 15:21:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:13.219 15:21:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:13.219 15:21:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:13.219 15:21:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:13.219 15:21:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:13.219 15:21:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:13.219 15:21:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:13.219 15:21:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:13.219 15:21:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:13.479 15:21:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:13.479 "name": "raid_bdev1", 00:15:13.479 "uuid": "d41199a8-b2e7-484a-a502-e4fec162c91e", 00:15:13.479 "strip_size_kb": 64, 00:15:13.479 "state": "online", 00:15:13.479 "raid_level": "raid5f", 00:15:13.479 "superblock": false, 00:15:13.479 "num_base_bdevs": 3, 00:15:13.479 "num_base_bdevs_discovered": 3, 00:15:13.479 "num_base_bdevs_operational": 3, 00:15:13.479 "base_bdevs_list": [ 00:15:13.479 { 00:15:13.479 "name": "spare", 00:15:13.479 "uuid": "ae650596-5d7a-5db8-9c0b-c2e81660fafd", 00:15:13.479 "is_configured": true, 00:15:13.479 "data_offset": 0, 00:15:13.479 "data_size": 65536 00:15:13.479 }, 00:15:13.479 { 00:15:13.479 "name": "BaseBdev2", 00:15:13.479 "uuid": "0ae5b1b0-1bb6-53af-83e2-1c19cd05b07b", 00:15:13.479 "is_configured": true, 00:15:13.479 "data_offset": 0, 00:15:13.479 "data_size": 65536 00:15:13.479 }, 00:15:13.479 { 00:15:13.479 "name": "BaseBdev3", 00:15:13.479 "uuid": "a64b1081-9e6a-5aba-adb8-dd4b0b2a2092", 00:15:13.479 "is_configured": true, 00:15:13.479 "data_offset": 0, 00:15:13.479 "data_size": 65536 00:15:13.479 } 00:15:13.479 ] 00:15:13.479 }' 00:15:13.479 15:21:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:13.479 15:21:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:13.479 15:21:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:13.479 15:21:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:13.479 15:21:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:13.479 15:21:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:13.479 15:21:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:13.479 15:21:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:13.479 15:21:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:13.479 15:21:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:13.479 15:21:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:13.479 15:21:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:13.479 15:21:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:13.479 15:21:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:13.479 15:21:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:13.479 15:21:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:13.479 15:21:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:13.479 15:21:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:13.479 15:21:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:13.479 15:21:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:13.479 "name": "raid_bdev1", 00:15:13.480 "uuid": "d41199a8-b2e7-484a-a502-e4fec162c91e", 00:15:13.480 "strip_size_kb": 64, 00:15:13.480 "state": "online", 00:15:13.480 "raid_level": "raid5f", 00:15:13.480 "superblock": false, 00:15:13.480 "num_base_bdevs": 3, 00:15:13.480 "num_base_bdevs_discovered": 3, 00:15:13.480 "num_base_bdevs_operational": 3, 00:15:13.480 "base_bdevs_list": [ 00:15:13.480 { 00:15:13.480 "name": "spare", 00:15:13.480 "uuid": "ae650596-5d7a-5db8-9c0b-c2e81660fafd", 00:15:13.480 "is_configured": true, 00:15:13.480 "data_offset": 0, 00:15:13.480 "data_size": 65536 00:15:13.480 }, 00:15:13.480 { 00:15:13.480 "name": "BaseBdev2", 00:15:13.480 "uuid": "0ae5b1b0-1bb6-53af-83e2-1c19cd05b07b", 00:15:13.480 "is_configured": true, 00:15:13.480 "data_offset": 0, 00:15:13.480 "data_size": 65536 00:15:13.480 }, 00:15:13.480 { 00:15:13.480 "name": "BaseBdev3", 00:15:13.480 "uuid": "a64b1081-9e6a-5aba-adb8-dd4b0b2a2092", 00:15:13.480 "is_configured": true, 00:15:13.480 "data_offset": 0, 00:15:13.480 "data_size": 65536 00:15:13.480 } 00:15:13.480 ] 00:15:13.480 }' 00:15:13.480 15:21:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:13.480 15:21:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:13.740 15:21:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:13.740 15:21:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:13.740 15:21:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:13.740 [2024-11-19 15:21:04.065266] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:13.740 [2024-11-19 15:21:04.065293] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:13.740 [2024-11-19 15:21:04.065370] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:13.740 [2024-11-19 15:21:04.065438] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:13.740 [2024-11-19 15:21:04.065447] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:15:13.740 15:21:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:13.740 15:21:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:13.740 15:21:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:13.740 15:21:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:13.740 15:21:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:15:14.000 15:21:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:14.000 15:21:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:15:14.000 15:21:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:15:14.000 15:21:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:15:14.000 15:21:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:15:14.000 15:21:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:15:14.000 15:21:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:15:14.000 15:21:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:15:14.000 15:21:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:15:14.000 15:21:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:15:14.000 15:21:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:15:14.000 15:21:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:15:14.000 15:21:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:14.000 15:21:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:15:14.000 /dev/nbd0 00:15:14.260 15:21:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:15:14.260 15:21:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:15:14.260 15:21:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:15:14.260 15:21:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:15:14.260 15:21:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:15:14.260 15:21:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:15:14.260 15:21:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:15:14.260 15:21:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@877 -- # break 00:15:14.260 15:21:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:15:14.261 15:21:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:15:14.261 15:21:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:14.261 1+0 records in 00:15:14.261 1+0 records out 00:15:14.261 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000538382 s, 7.6 MB/s 00:15:14.261 15:21:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:14.261 15:21:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:15:14.261 15:21:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:14.261 15:21:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:15:14.261 15:21:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:15:14.261 15:21:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:14.261 15:21:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:14.261 15:21:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:15:14.261 /dev/nbd1 00:15:14.521 15:21:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:15:14.521 15:21:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:15:14.521 15:21:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:15:14.521 15:21:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:15:14.521 15:21:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:15:14.521 15:21:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:15:14.521 15:21:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:15:14.521 15:21:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@877 -- # break 00:15:14.521 15:21:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:15:14.521 15:21:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:15:14.521 15:21:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:14.521 1+0 records in 00:15:14.521 1+0 records out 00:15:14.521 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000531064 s, 7.7 MB/s 00:15:14.521 15:21:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:14.521 15:21:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:15:14.521 15:21:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:14.521 15:21:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:15:14.521 15:21:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:15:14.521 15:21:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:14.521 15:21:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:14.521 15:21:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:15:14.521 15:21:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:15:14.521 15:21:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:15:14.521 15:21:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:15:14.521 15:21:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:15:14.521 15:21:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:15:14.521 15:21:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:14.521 15:21:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:15:14.781 15:21:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:15:14.781 15:21:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:15:14.781 15:21:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:15:14.781 15:21:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:14.781 15:21:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:14.781 15:21:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:15:14.781 15:21:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:15:14.781 15:21:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:15:14.781 15:21:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:14.781 15:21:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:15:15.042 15:21:05 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:15:15.042 15:21:05 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:15:15.042 15:21:05 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:15:15.042 15:21:05 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:15.042 15:21:05 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:15.042 15:21:05 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:15:15.042 15:21:05 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:15:15.042 15:21:05 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:15:15.042 15:21:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:15:15.042 15:21:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 92099 00:15:15.042 15:21:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@954 -- # '[' -z 92099 ']' 00:15:15.042 15:21:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@958 -- # kill -0 92099 00:15:15.042 15:21:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@959 -- # uname 00:15:15.042 15:21:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:15:15.042 15:21:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 92099 00:15:15.042 15:21:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:15:15.042 killing process with pid 92099 00:15:15.042 Received shutdown signal, test time was about 60.000000 seconds 00:15:15.042 00:15:15.042 Latency(us) 00:15:15.042 [2024-11-19T15:21:05.381Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:15:15.042 [2024-11-19T15:21:05.381Z] =================================================================================================================== 00:15:15.042 [2024-11-19T15:21:05.381Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:15:15.042 15:21:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:15:15.042 15:21:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 92099' 00:15:15.042 15:21:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@973 -- # kill 92099 00:15:15.042 [2024-11-19 15:21:05.202559] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:15.042 15:21:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@978 -- # wait 92099 00:15:15.042 [2024-11-19 15:21:05.243628] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:15.303 15:21:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:15:15.303 00:15:15.303 real 0m13.777s 00:15:15.303 user 0m17.335s 00:15:15.303 sys 0m2.037s 00:15:15.303 15:21:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:15:15.303 ************************************ 00:15:15.303 END TEST raid5f_rebuild_test 00:15:15.303 ************************************ 00:15:15.303 15:21:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:15.303 15:21:05 bdev_raid -- bdev/bdev_raid.sh@991 -- # run_test raid5f_rebuild_test_sb raid_rebuild_test raid5f 3 true false true 00:15:15.303 15:21:05 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:15:15.303 15:21:05 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:15:15.303 15:21:05 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:15.303 ************************************ 00:15:15.303 START TEST raid5f_rebuild_test_sb 00:15:15.303 ************************************ 00:15:15.303 15:21:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid5f 3 true false true 00:15:15.303 15:21:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:15:15.303 15:21:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=3 00:15:15.303 15:21:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:15:15.303 15:21:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:15:15.303 15:21:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:15:15.303 15:21:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:15:15.303 15:21:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:15.303 15:21:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:15:15.303 15:21:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:15.303 15:21:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:15.303 15:21:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:15:15.303 15:21:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:15.303 15:21:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:15.303 15:21:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:15:15.303 15:21:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:15.303 15:21:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:15.303 15:21:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:15:15.303 15:21:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:15:15.303 15:21:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:15:15.303 15:21:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:15:15.303 15:21:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:15:15.303 15:21:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:15:15.303 15:21:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:15:15.303 15:21:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:15:15.303 15:21:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:15:15.303 15:21:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:15:15.303 15:21:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:15:15.303 15:21:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:15:15.303 15:21:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:15:15.303 15:21:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=92522 00:15:15.303 15:21:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:15:15.303 15:21:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 92522 00:15:15.303 15:21:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@835 -- # '[' -z 92522 ']' 00:15:15.303 15:21:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:15.303 15:21:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:15:15.303 15:21:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:15.303 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:15.303 15:21:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:15:15.303 15:21:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:15.303 [2024-11-19 15:21:05.628516] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:15:15.303 I/O size of 3145728 is greater than zero copy threshold (65536). 00:15:15.303 Zero copy mechanism will not be used. 00:15:15.303 [2024-11-19 15:21:05.628733] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid92522 ] 00:15:15.563 [2024-11-19 15:21:05.785887] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:15.563 [2024-11-19 15:21:05.810612] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:15.563 [2024-11-19 15:21:05.854205] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:15.563 [2024-11-19 15:21:05.854239] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:16.133 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:15:16.133 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@868 -- # return 0 00:15:16.133 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:16.133 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:15:16.133 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:16.133 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:16.133 BaseBdev1_malloc 00:15:16.133 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:16.133 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:15:16.133 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:16.133 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:16.133 [2024-11-19 15:21:06.457032] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:15:16.133 [2024-11-19 15:21:06.457145] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:16.133 [2024-11-19 15:21:06.457180] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:15:16.133 [2024-11-19 15:21:06.457199] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:16.133 [2024-11-19 15:21:06.459292] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:16.133 [2024-11-19 15:21:06.459347] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:15:16.133 BaseBdev1 00:15:16.133 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:16.133 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:16.133 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:15:16.133 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:16.133 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:16.394 BaseBdev2_malloc 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:16.394 [2024-11-19 15:21:06.485545] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:15:16.394 [2024-11-19 15:21:06.485589] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:16.394 [2024-11-19 15:21:06.485607] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:15:16.394 [2024-11-19 15:21:06.485615] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:16.394 [2024-11-19 15:21:06.487632] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:16.394 [2024-11-19 15:21:06.487672] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:15:16.394 BaseBdev2 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:16.394 BaseBdev3_malloc 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:16.394 [2024-11-19 15:21:06.514121] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:15:16.394 [2024-11-19 15:21:06.514171] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:16.394 [2024-11-19 15:21:06.514191] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:15:16.394 [2024-11-19 15:21:06.514201] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:16.394 [2024-11-19 15:21:06.516228] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:16.394 [2024-11-19 15:21:06.516274] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:15:16.394 BaseBdev3 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:16.394 spare_malloc 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:16.394 spare_delay 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:16.394 [2024-11-19 15:21:06.572070] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:16.394 [2024-11-19 15:21:06.572149] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:16.394 [2024-11-19 15:21:06.572193] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:15:16.394 [2024-11-19 15:21:06.572210] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:16.394 [2024-11-19 15:21:06.574945] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:16.394 [2024-11-19 15:21:06.575000] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:16.394 spare 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:16.394 [2024-11-19 15:21:06.584075] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:16.394 [2024-11-19 15:21:06.585865] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:16.394 [2024-11-19 15:21:06.585927] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:15:16.394 [2024-11-19 15:21:06.586153] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:15:16.394 [2024-11-19 15:21:06.586201] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:15:16.394 [2024-11-19 15:21:06.586452] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:15:16.394 [2024-11-19 15:21:06.586908] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:15:16.394 [2024-11-19 15:21:06.586930] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:15:16.394 [2024-11-19 15:21:06.587062] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:16.394 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:16.394 "name": "raid_bdev1", 00:15:16.394 "uuid": "ff3a5741-c604-4f3c-98f3-f3eb4f28c032", 00:15:16.394 "strip_size_kb": 64, 00:15:16.394 "state": "online", 00:15:16.394 "raid_level": "raid5f", 00:15:16.394 "superblock": true, 00:15:16.394 "num_base_bdevs": 3, 00:15:16.394 "num_base_bdevs_discovered": 3, 00:15:16.394 "num_base_bdevs_operational": 3, 00:15:16.394 "base_bdevs_list": [ 00:15:16.394 { 00:15:16.394 "name": "BaseBdev1", 00:15:16.394 "uuid": "63c068eb-bbea-5faf-a6a8-b81b81565d09", 00:15:16.394 "is_configured": true, 00:15:16.395 "data_offset": 2048, 00:15:16.395 "data_size": 63488 00:15:16.395 }, 00:15:16.395 { 00:15:16.395 "name": "BaseBdev2", 00:15:16.395 "uuid": "71014908-0916-5fee-85d5-2a2e4a4b4bc5", 00:15:16.395 "is_configured": true, 00:15:16.395 "data_offset": 2048, 00:15:16.395 "data_size": 63488 00:15:16.395 }, 00:15:16.395 { 00:15:16.395 "name": "BaseBdev3", 00:15:16.395 "uuid": "1cb48daf-e0ed-56fd-8811-afa342721340", 00:15:16.395 "is_configured": true, 00:15:16.395 "data_offset": 2048, 00:15:16.395 "data_size": 63488 00:15:16.395 } 00:15:16.395 ] 00:15:16.395 }' 00:15:16.395 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:16.395 15:21:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:16.965 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:16.965 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:15:16.965 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:16.965 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:16.965 [2024-11-19 15:21:07.028031] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:16.965 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:16.965 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=126976 00:15:16.965 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:16.965 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:16.965 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:16.965 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:15:16.965 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:16.965 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:15:16.965 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:15:16.965 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:15:16.965 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:15:16.965 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:15:16.965 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:15:16.965 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:15:16.965 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:15:16.965 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:15:16.965 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:15:16.965 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:15:16.965 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:15:16.965 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:15:16.965 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:15:16.965 [2024-11-19 15:21:07.287412] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:15:17.225 /dev/nbd0 00:15:17.225 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:15:17.225 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:15:17.225 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:15:17.225 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:15:17.225 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:15:17.225 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:15:17.225 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:15:17.225 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:15:17.225 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:15:17.225 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:15:17.225 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:17.225 1+0 records in 00:15:17.225 1+0 records out 00:15:17.225 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000497541 s, 8.2 MB/s 00:15:17.225 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:17.225 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:15:17.225 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:17.225 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:15:17.225 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:15:17.225 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:17.226 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:15:17.226 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:15:17.226 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@630 -- # write_unit_size=256 00:15:17.226 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@631 -- # echo 128 00:15:17.226 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=131072 count=496 oflag=direct 00:15:17.486 496+0 records in 00:15:17.486 496+0 records out 00:15:17.486 65011712 bytes (65 MB, 62 MiB) copied, 0.302123 s, 215 MB/s 00:15:17.486 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:15:17.486 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:15:17.486 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:15:17.486 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:15:17.486 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:15:17.486 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:17.486 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:15:17.746 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:15:17.746 [2024-11-19 15:21:07.880996] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:17.746 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:15:17.746 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:15:17.746 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:17.746 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:17.746 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:15:17.746 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:15:17.746 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:15:17.746 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:15:17.746 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:17.746 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:17.746 [2024-11-19 15:21:07.897041] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:17.746 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:17.746 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:15:17.746 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:17.746 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:17.746 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:17.746 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:17.746 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:17.746 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:17.746 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:17.746 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:17.746 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:17.746 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:17.746 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:17.746 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:17.746 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:17.746 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:17.746 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:17.746 "name": "raid_bdev1", 00:15:17.746 "uuid": "ff3a5741-c604-4f3c-98f3-f3eb4f28c032", 00:15:17.746 "strip_size_kb": 64, 00:15:17.746 "state": "online", 00:15:17.746 "raid_level": "raid5f", 00:15:17.746 "superblock": true, 00:15:17.746 "num_base_bdevs": 3, 00:15:17.746 "num_base_bdevs_discovered": 2, 00:15:17.746 "num_base_bdevs_operational": 2, 00:15:17.746 "base_bdevs_list": [ 00:15:17.746 { 00:15:17.746 "name": null, 00:15:17.746 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:17.746 "is_configured": false, 00:15:17.746 "data_offset": 0, 00:15:17.746 "data_size": 63488 00:15:17.746 }, 00:15:17.746 { 00:15:17.746 "name": "BaseBdev2", 00:15:17.746 "uuid": "71014908-0916-5fee-85d5-2a2e4a4b4bc5", 00:15:17.746 "is_configured": true, 00:15:17.746 "data_offset": 2048, 00:15:17.746 "data_size": 63488 00:15:17.746 }, 00:15:17.746 { 00:15:17.746 "name": "BaseBdev3", 00:15:17.746 "uuid": "1cb48daf-e0ed-56fd-8811-afa342721340", 00:15:17.746 "is_configured": true, 00:15:17.746 "data_offset": 2048, 00:15:17.746 "data_size": 63488 00:15:17.746 } 00:15:17.746 ] 00:15:17.746 }' 00:15:17.746 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:17.746 15:21:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:18.006 15:21:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:18.006 15:21:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:18.006 15:21:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:18.006 [2024-11-19 15:21:08.316304] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:18.006 [2024-11-19 15:21:08.320906] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000255d0 00:15:18.006 15:21:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:18.006 15:21:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:15:18.006 [2024-11-19 15:21:08.323110] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:19.388 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:19.388 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:19.388 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:19.388 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:19.388 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:19.388 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:19.388 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:19.388 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:19.388 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:19.388 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:19.388 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:19.388 "name": "raid_bdev1", 00:15:19.388 "uuid": "ff3a5741-c604-4f3c-98f3-f3eb4f28c032", 00:15:19.388 "strip_size_kb": 64, 00:15:19.388 "state": "online", 00:15:19.388 "raid_level": "raid5f", 00:15:19.388 "superblock": true, 00:15:19.388 "num_base_bdevs": 3, 00:15:19.388 "num_base_bdevs_discovered": 3, 00:15:19.388 "num_base_bdevs_operational": 3, 00:15:19.388 "process": { 00:15:19.388 "type": "rebuild", 00:15:19.388 "target": "spare", 00:15:19.388 "progress": { 00:15:19.388 "blocks": 20480, 00:15:19.388 "percent": 16 00:15:19.388 } 00:15:19.388 }, 00:15:19.388 "base_bdevs_list": [ 00:15:19.388 { 00:15:19.388 "name": "spare", 00:15:19.388 "uuid": "78e4b36a-286b-56f2-9b2c-22d5b63f416e", 00:15:19.388 "is_configured": true, 00:15:19.388 "data_offset": 2048, 00:15:19.388 "data_size": 63488 00:15:19.388 }, 00:15:19.388 { 00:15:19.388 "name": "BaseBdev2", 00:15:19.388 "uuid": "71014908-0916-5fee-85d5-2a2e4a4b4bc5", 00:15:19.388 "is_configured": true, 00:15:19.388 "data_offset": 2048, 00:15:19.388 "data_size": 63488 00:15:19.388 }, 00:15:19.388 { 00:15:19.388 "name": "BaseBdev3", 00:15:19.388 "uuid": "1cb48daf-e0ed-56fd-8811-afa342721340", 00:15:19.388 "is_configured": true, 00:15:19.388 "data_offset": 2048, 00:15:19.388 "data_size": 63488 00:15:19.388 } 00:15:19.388 ] 00:15:19.388 }' 00:15:19.388 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:19.388 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:19.388 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:19.388 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:19.388 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:15:19.388 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:19.388 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:19.388 [2024-11-19 15:21:09.486990] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:19.388 [2024-11-19 15:21:09.529496] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:19.388 [2024-11-19 15:21:09.529599] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:19.388 [2024-11-19 15:21:09.529616] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:19.388 [2024-11-19 15:21:09.529635] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:19.388 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:19.388 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:15:19.388 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:19.388 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:19.388 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:19.388 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:19.388 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:19.388 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:19.388 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:19.388 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:19.388 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:19.388 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:19.388 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:19.388 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:19.388 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:19.388 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:19.388 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:19.388 "name": "raid_bdev1", 00:15:19.388 "uuid": "ff3a5741-c604-4f3c-98f3-f3eb4f28c032", 00:15:19.388 "strip_size_kb": 64, 00:15:19.388 "state": "online", 00:15:19.388 "raid_level": "raid5f", 00:15:19.388 "superblock": true, 00:15:19.388 "num_base_bdevs": 3, 00:15:19.388 "num_base_bdevs_discovered": 2, 00:15:19.388 "num_base_bdevs_operational": 2, 00:15:19.388 "base_bdevs_list": [ 00:15:19.388 { 00:15:19.388 "name": null, 00:15:19.388 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:19.388 "is_configured": false, 00:15:19.388 "data_offset": 0, 00:15:19.388 "data_size": 63488 00:15:19.388 }, 00:15:19.388 { 00:15:19.388 "name": "BaseBdev2", 00:15:19.388 "uuid": "71014908-0916-5fee-85d5-2a2e4a4b4bc5", 00:15:19.388 "is_configured": true, 00:15:19.388 "data_offset": 2048, 00:15:19.388 "data_size": 63488 00:15:19.388 }, 00:15:19.388 { 00:15:19.388 "name": "BaseBdev3", 00:15:19.388 "uuid": "1cb48daf-e0ed-56fd-8811-afa342721340", 00:15:19.388 "is_configured": true, 00:15:19.388 "data_offset": 2048, 00:15:19.388 "data_size": 63488 00:15:19.388 } 00:15:19.388 ] 00:15:19.388 }' 00:15:19.388 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:19.388 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:19.648 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:19.648 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:19.648 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:19.648 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:19.649 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:19.649 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:19.649 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:19.649 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:19.649 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:19.649 15:21:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:19.909 15:21:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:19.909 "name": "raid_bdev1", 00:15:19.909 "uuid": "ff3a5741-c604-4f3c-98f3-f3eb4f28c032", 00:15:19.909 "strip_size_kb": 64, 00:15:19.909 "state": "online", 00:15:19.909 "raid_level": "raid5f", 00:15:19.909 "superblock": true, 00:15:19.909 "num_base_bdevs": 3, 00:15:19.909 "num_base_bdevs_discovered": 2, 00:15:19.909 "num_base_bdevs_operational": 2, 00:15:19.909 "base_bdevs_list": [ 00:15:19.909 { 00:15:19.909 "name": null, 00:15:19.909 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:19.909 "is_configured": false, 00:15:19.909 "data_offset": 0, 00:15:19.909 "data_size": 63488 00:15:19.909 }, 00:15:19.909 { 00:15:19.909 "name": "BaseBdev2", 00:15:19.909 "uuid": "71014908-0916-5fee-85d5-2a2e4a4b4bc5", 00:15:19.909 "is_configured": true, 00:15:19.909 "data_offset": 2048, 00:15:19.909 "data_size": 63488 00:15:19.909 }, 00:15:19.909 { 00:15:19.909 "name": "BaseBdev3", 00:15:19.909 "uuid": "1cb48daf-e0ed-56fd-8811-afa342721340", 00:15:19.909 "is_configured": true, 00:15:19.909 "data_offset": 2048, 00:15:19.909 "data_size": 63488 00:15:19.909 } 00:15:19.909 ] 00:15:19.909 }' 00:15:19.909 15:21:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:19.909 15:21:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:19.909 15:21:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:19.909 15:21:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:19.909 15:21:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:19.909 15:21:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:19.909 15:21:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:19.909 [2024-11-19 15:21:10.106299] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:19.909 [2024-11-19 15:21:10.110009] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000256a0 00:15:19.909 [2024-11-19 15:21:10.112106] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:19.909 15:21:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:19.909 15:21:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:15:20.849 15:21:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:20.849 15:21:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:20.849 15:21:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:20.849 15:21:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:20.849 15:21:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:20.849 15:21:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:20.849 15:21:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:20.849 15:21:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:20.849 15:21:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:20.849 15:21:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:20.849 15:21:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:20.849 "name": "raid_bdev1", 00:15:20.849 "uuid": "ff3a5741-c604-4f3c-98f3-f3eb4f28c032", 00:15:20.849 "strip_size_kb": 64, 00:15:20.849 "state": "online", 00:15:20.849 "raid_level": "raid5f", 00:15:20.849 "superblock": true, 00:15:20.849 "num_base_bdevs": 3, 00:15:20.849 "num_base_bdevs_discovered": 3, 00:15:20.849 "num_base_bdevs_operational": 3, 00:15:20.849 "process": { 00:15:20.849 "type": "rebuild", 00:15:20.849 "target": "spare", 00:15:20.849 "progress": { 00:15:20.849 "blocks": 20480, 00:15:20.849 "percent": 16 00:15:20.849 } 00:15:20.849 }, 00:15:20.849 "base_bdevs_list": [ 00:15:20.849 { 00:15:20.849 "name": "spare", 00:15:20.849 "uuid": "78e4b36a-286b-56f2-9b2c-22d5b63f416e", 00:15:20.849 "is_configured": true, 00:15:20.849 "data_offset": 2048, 00:15:20.849 "data_size": 63488 00:15:20.849 }, 00:15:20.849 { 00:15:20.849 "name": "BaseBdev2", 00:15:20.849 "uuid": "71014908-0916-5fee-85d5-2a2e4a4b4bc5", 00:15:20.849 "is_configured": true, 00:15:20.849 "data_offset": 2048, 00:15:20.849 "data_size": 63488 00:15:20.849 }, 00:15:20.849 { 00:15:20.849 "name": "BaseBdev3", 00:15:20.849 "uuid": "1cb48daf-e0ed-56fd-8811-afa342721340", 00:15:20.849 "is_configured": true, 00:15:20.849 "data_offset": 2048, 00:15:20.849 "data_size": 63488 00:15:20.849 } 00:15:20.849 ] 00:15:20.849 }' 00:15:20.849 15:21:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:21.109 15:21:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:21.109 15:21:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:21.109 15:21:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:21.109 15:21:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:15:21.109 15:21:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:15:21.109 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:15:21.109 15:21:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=3 00:15:21.109 15:21:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:15:21.109 15:21:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=469 00:15:21.109 15:21:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:21.109 15:21:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:21.109 15:21:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:21.109 15:21:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:21.109 15:21:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:21.109 15:21:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:21.110 15:21:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:21.110 15:21:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:21.110 15:21:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:21.110 15:21:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:21.110 15:21:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:21.110 15:21:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:21.110 "name": "raid_bdev1", 00:15:21.110 "uuid": "ff3a5741-c604-4f3c-98f3-f3eb4f28c032", 00:15:21.110 "strip_size_kb": 64, 00:15:21.110 "state": "online", 00:15:21.110 "raid_level": "raid5f", 00:15:21.110 "superblock": true, 00:15:21.110 "num_base_bdevs": 3, 00:15:21.110 "num_base_bdevs_discovered": 3, 00:15:21.110 "num_base_bdevs_operational": 3, 00:15:21.110 "process": { 00:15:21.110 "type": "rebuild", 00:15:21.110 "target": "spare", 00:15:21.110 "progress": { 00:15:21.110 "blocks": 22528, 00:15:21.110 "percent": 17 00:15:21.110 } 00:15:21.110 }, 00:15:21.110 "base_bdevs_list": [ 00:15:21.110 { 00:15:21.110 "name": "spare", 00:15:21.110 "uuid": "78e4b36a-286b-56f2-9b2c-22d5b63f416e", 00:15:21.110 "is_configured": true, 00:15:21.110 "data_offset": 2048, 00:15:21.110 "data_size": 63488 00:15:21.110 }, 00:15:21.110 { 00:15:21.110 "name": "BaseBdev2", 00:15:21.110 "uuid": "71014908-0916-5fee-85d5-2a2e4a4b4bc5", 00:15:21.110 "is_configured": true, 00:15:21.110 "data_offset": 2048, 00:15:21.110 "data_size": 63488 00:15:21.110 }, 00:15:21.110 { 00:15:21.110 "name": "BaseBdev3", 00:15:21.110 "uuid": "1cb48daf-e0ed-56fd-8811-afa342721340", 00:15:21.110 "is_configured": true, 00:15:21.110 "data_offset": 2048, 00:15:21.110 "data_size": 63488 00:15:21.110 } 00:15:21.110 ] 00:15:21.110 }' 00:15:21.110 15:21:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:21.110 15:21:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:21.110 15:21:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:21.110 15:21:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:21.110 15:21:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:22.123 15:21:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:22.123 15:21:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:22.123 15:21:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:22.123 15:21:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:22.123 15:21:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:22.123 15:21:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:22.123 15:21:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:22.123 15:21:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:22.123 15:21:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:22.123 15:21:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:22.123 15:21:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:22.123 15:21:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:22.123 "name": "raid_bdev1", 00:15:22.123 "uuid": "ff3a5741-c604-4f3c-98f3-f3eb4f28c032", 00:15:22.123 "strip_size_kb": 64, 00:15:22.123 "state": "online", 00:15:22.123 "raid_level": "raid5f", 00:15:22.123 "superblock": true, 00:15:22.123 "num_base_bdevs": 3, 00:15:22.123 "num_base_bdevs_discovered": 3, 00:15:22.123 "num_base_bdevs_operational": 3, 00:15:22.124 "process": { 00:15:22.124 "type": "rebuild", 00:15:22.124 "target": "spare", 00:15:22.124 "progress": { 00:15:22.124 "blocks": 45056, 00:15:22.124 "percent": 35 00:15:22.124 } 00:15:22.124 }, 00:15:22.124 "base_bdevs_list": [ 00:15:22.124 { 00:15:22.124 "name": "spare", 00:15:22.124 "uuid": "78e4b36a-286b-56f2-9b2c-22d5b63f416e", 00:15:22.124 "is_configured": true, 00:15:22.124 "data_offset": 2048, 00:15:22.124 "data_size": 63488 00:15:22.124 }, 00:15:22.124 { 00:15:22.124 "name": "BaseBdev2", 00:15:22.124 "uuid": "71014908-0916-5fee-85d5-2a2e4a4b4bc5", 00:15:22.124 "is_configured": true, 00:15:22.124 "data_offset": 2048, 00:15:22.124 "data_size": 63488 00:15:22.124 }, 00:15:22.124 { 00:15:22.124 "name": "BaseBdev3", 00:15:22.124 "uuid": "1cb48daf-e0ed-56fd-8811-afa342721340", 00:15:22.124 "is_configured": true, 00:15:22.124 "data_offset": 2048, 00:15:22.124 "data_size": 63488 00:15:22.124 } 00:15:22.124 ] 00:15:22.124 }' 00:15:22.124 15:21:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:22.384 15:21:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:22.384 15:21:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:22.384 15:21:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:22.384 15:21:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:23.325 15:21:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:23.325 15:21:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:23.325 15:21:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:23.325 15:21:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:23.325 15:21:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:23.325 15:21:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:23.325 15:21:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:23.325 15:21:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:23.325 15:21:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:23.325 15:21:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:23.325 15:21:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:23.325 15:21:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:23.325 "name": "raid_bdev1", 00:15:23.325 "uuid": "ff3a5741-c604-4f3c-98f3-f3eb4f28c032", 00:15:23.325 "strip_size_kb": 64, 00:15:23.325 "state": "online", 00:15:23.325 "raid_level": "raid5f", 00:15:23.325 "superblock": true, 00:15:23.325 "num_base_bdevs": 3, 00:15:23.325 "num_base_bdevs_discovered": 3, 00:15:23.325 "num_base_bdevs_operational": 3, 00:15:23.325 "process": { 00:15:23.325 "type": "rebuild", 00:15:23.325 "target": "spare", 00:15:23.325 "progress": { 00:15:23.325 "blocks": 69632, 00:15:23.325 "percent": 54 00:15:23.325 } 00:15:23.325 }, 00:15:23.325 "base_bdevs_list": [ 00:15:23.325 { 00:15:23.325 "name": "spare", 00:15:23.325 "uuid": "78e4b36a-286b-56f2-9b2c-22d5b63f416e", 00:15:23.325 "is_configured": true, 00:15:23.325 "data_offset": 2048, 00:15:23.325 "data_size": 63488 00:15:23.325 }, 00:15:23.325 { 00:15:23.325 "name": "BaseBdev2", 00:15:23.325 "uuid": "71014908-0916-5fee-85d5-2a2e4a4b4bc5", 00:15:23.325 "is_configured": true, 00:15:23.325 "data_offset": 2048, 00:15:23.325 "data_size": 63488 00:15:23.325 }, 00:15:23.325 { 00:15:23.325 "name": "BaseBdev3", 00:15:23.325 "uuid": "1cb48daf-e0ed-56fd-8811-afa342721340", 00:15:23.325 "is_configured": true, 00:15:23.325 "data_offset": 2048, 00:15:23.325 "data_size": 63488 00:15:23.325 } 00:15:23.325 ] 00:15:23.325 }' 00:15:23.325 15:21:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:23.325 15:21:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:23.325 15:21:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:23.586 15:21:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:23.586 15:21:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:24.526 15:21:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:24.526 15:21:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:24.526 15:21:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:24.526 15:21:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:24.526 15:21:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:24.526 15:21:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:24.526 15:21:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:24.526 15:21:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:24.526 15:21:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:24.526 15:21:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:24.526 15:21:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:24.526 15:21:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:24.526 "name": "raid_bdev1", 00:15:24.526 "uuid": "ff3a5741-c604-4f3c-98f3-f3eb4f28c032", 00:15:24.526 "strip_size_kb": 64, 00:15:24.526 "state": "online", 00:15:24.526 "raid_level": "raid5f", 00:15:24.526 "superblock": true, 00:15:24.526 "num_base_bdevs": 3, 00:15:24.527 "num_base_bdevs_discovered": 3, 00:15:24.527 "num_base_bdevs_operational": 3, 00:15:24.527 "process": { 00:15:24.527 "type": "rebuild", 00:15:24.527 "target": "spare", 00:15:24.527 "progress": { 00:15:24.527 "blocks": 92160, 00:15:24.527 "percent": 72 00:15:24.527 } 00:15:24.527 }, 00:15:24.527 "base_bdevs_list": [ 00:15:24.527 { 00:15:24.527 "name": "spare", 00:15:24.527 "uuid": "78e4b36a-286b-56f2-9b2c-22d5b63f416e", 00:15:24.527 "is_configured": true, 00:15:24.527 "data_offset": 2048, 00:15:24.527 "data_size": 63488 00:15:24.527 }, 00:15:24.527 { 00:15:24.527 "name": "BaseBdev2", 00:15:24.527 "uuid": "71014908-0916-5fee-85d5-2a2e4a4b4bc5", 00:15:24.527 "is_configured": true, 00:15:24.527 "data_offset": 2048, 00:15:24.527 "data_size": 63488 00:15:24.527 }, 00:15:24.527 { 00:15:24.527 "name": "BaseBdev3", 00:15:24.527 "uuid": "1cb48daf-e0ed-56fd-8811-afa342721340", 00:15:24.527 "is_configured": true, 00:15:24.527 "data_offset": 2048, 00:15:24.527 "data_size": 63488 00:15:24.527 } 00:15:24.527 ] 00:15:24.527 }' 00:15:24.527 15:21:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:24.527 15:21:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:24.527 15:21:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:24.527 15:21:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:24.527 15:21:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:25.910 15:21:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:25.910 15:21:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:25.910 15:21:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:25.910 15:21:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:25.910 15:21:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:25.910 15:21:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:25.910 15:21:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:25.910 15:21:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:25.910 15:21:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:25.910 15:21:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:25.910 15:21:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:25.910 15:21:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:25.910 "name": "raid_bdev1", 00:15:25.910 "uuid": "ff3a5741-c604-4f3c-98f3-f3eb4f28c032", 00:15:25.910 "strip_size_kb": 64, 00:15:25.910 "state": "online", 00:15:25.910 "raid_level": "raid5f", 00:15:25.910 "superblock": true, 00:15:25.910 "num_base_bdevs": 3, 00:15:25.910 "num_base_bdevs_discovered": 3, 00:15:25.910 "num_base_bdevs_operational": 3, 00:15:25.910 "process": { 00:15:25.910 "type": "rebuild", 00:15:25.910 "target": "spare", 00:15:25.910 "progress": { 00:15:25.910 "blocks": 116736, 00:15:25.910 "percent": 91 00:15:25.910 } 00:15:25.910 }, 00:15:25.910 "base_bdevs_list": [ 00:15:25.910 { 00:15:25.910 "name": "spare", 00:15:25.910 "uuid": "78e4b36a-286b-56f2-9b2c-22d5b63f416e", 00:15:25.910 "is_configured": true, 00:15:25.910 "data_offset": 2048, 00:15:25.910 "data_size": 63488 00:15:25.910 }, 00:15:25.910 { 00:15:25.910 "name": "BaseBdev2", 00:15:25.910 "uuid": "71014908-0916-5fee-85d5-2a2e4a4b4bc5", 00:15:25.910 "is_configured": true, 00:15:25.910 "data_offset": 2048, 00:15:25.910 "data_size": 63488 00:15:25.910 }, 00:15:25.910 { 00:15:25.910 "name": "BaseBdev3", 00:15:25.910 "uuid": "1cb48daf-e0ed-56fd-8811-afa342721340", 00:15:25.910 "is_configured": true, 00:15:25.910 "data_offset": 2048, 00:15:25.910 "data_size": 63488 00:15:25.910 } 00:15:25.910 ] 00:15:25.910 }' 00:15:25.910 15:21:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:25.910 15:21:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:25.910 15:21:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:25.910 15:21:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:25.910 15:21:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:26.171 [2024-11-19 15:21:16.342861] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:15:26.171 [2024-11-19 15:21:16.343006] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:15:26.171 [2024-11-19 15:21:16.343143] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:26.741 15:21:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:26.741 15:21:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:26.741 15:21:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:26.741 15:21:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:26.741 15:21:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:26.741 15:21:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:26.741 15:21:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:26.742 15:21:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:26.742 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:26.742 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:26.742 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:26.742 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:26.742 "name": "raid_bdev1", 00:15:26.742 "uuid": "ff3a5741-c604-4f3c-98f3-f3eb4f28c032", 00:15:26.742 "strip_size_kb": 64, 00:15:26.742 "state": "online", 00:15:26.742 "raid_level": "raid5f", 00:15:26.742 "superblock": true, 00:15:26.742 "num_base_bdevs": 3, 00:15:26.742 "num_base_bdevs_discovered": 3, 00:15:26.742 "num_base_bdevs_operational": 3, 00:15:26.742 "base_bdevs_list": [ 00:15:26.742 { 00:15:26.742 "name": "spare", 00:15:26.742 "uuid": "78e4b36a-286b-56f2-9b2c-22d5b63f416e", 00:15:26.742 "is_configured": true, 00:15:26.742 "data_offset": 2048, 00:15:26.742 "data_size": 63488 00:15:26.742 }, 00:15:26.742 { 00:15:26.742 "name": "BaseBdev2", 00:15:26.742 "uuid": "71014908-0916-5fee-85d5-2a2e4a4b4bc5", 00:15:26.742 "is_configured": true, 00:15:26.742 "data_offset": 2048, 00:15:26.742 "data_size": 63488 00:15:26.742 }, 00:15:26.742 { 00:15:26.742 "name": "BaseBdev3", 00:15:26.742 "uuid": "1cb48daf-e0ed-56fd-8811-afa342721340", 00:15:26.742 "is_configured": true, 00:15:26.742 "data_offset": 2048, 00:15:26.742 "data_size": 63488 00:15:26.742 } 00:15:26.742 ] 00:15:26.742 }' 00:15:26.742 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:27.002 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:15:27.002 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:27.002 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:15:27.002 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:15:27.002 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:27.002 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:27.002 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:27.002 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:27.002 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:27.002 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:27.002 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:27.002 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:27.002 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:27.002 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:27.002 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:27.002 "name": "raid_bdev1", 00:15:27.002 "uuid": "ff3a5741-c604-4f3c-98f3-f3eb4f28c032", 00:15:27.002 "strip_size_kb": 64, 00:15:27.002 "state": "online", 00:15:27.002 "raid_level": "raid5f", 00:15:27.002 "superblock": true, 00:15:27.002 "num_base_bdevs": 3, 00:15:27.002 "num_base_bdevs_discovered": 3, 00:15:27.002 "num_base_bdevs_operational": 3, 00:15:27.002 "base_bdevs_list": [ 00:15:27.002 { 00:15:27.002 "name": "spare", 00:15:27.002 "uuid": "78e4b36a-286b-56f2-9b2c-22d5b63f416e", 00:15:27.002 "is_configured": true, 00:15:27.002 "data_offset": 2048, 00:15:27.002 "data_size": 63488 00:15:27.002 }, 00:15:27.002 { 00:15:27.002 "name": "BaseBdev2", 00:15:27.002 "uuid": "71014908-0916-5fee-85d5-2a2e4a4b4bc5", 00:15:27.002 "is_configured": true, 00:15:27.002 "data_offset": 2048, 00:15:27.002 "data_size": 63488 00:15:27.002 }, 00:15:27.002 { 00:15:27.002 "name": "BaseBdev3", 00:15:27.002 "uuid": "1cb48daf-e0ed-56fd-8811-afa342721340", 00:15:27.002 "is_configured": true, 00:15:27.002 "data_offset": 2048, 00:15:27.002 "data_size": 63488 00:15:27.002 } 00:15:27.002 ] 00:15:27.002 }' 00:15:27.002 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:27.002 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:27.002 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:27.002 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:27.002 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:27.002 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:27.002 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:27.002 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:27.002 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:27.002 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:27.002 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:27.002 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:27.002 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:27.002 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:27.002 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:27.002 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:27.002 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:27.002 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:27.002 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:27.262 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:27.262 "name": "raid_bdev1", 00:15:27.262 "uuid": "ff3a5741-c604-4f3c-98f3-f3eb4f28c032", 00:15:27.262 "strip_size_kb": 64, 00:15:27.262 "state": "online", 00:15:27.262 "raid_level": "raid5f", 00:15:27.262 "superblock": true, 00:15:27.262 "num_base_bdevs": 3, 00:15:27.262 "num_base_bdevs_discovered": 3, 00:15:27.262 "num_base_bdevs_operational": 3, 00:15:27.262 "base_bdevs_list": [ 00:15:27.262 { 00:15:27.262 "name": "spare", 00:15:27.262 "uuid": "78e4b36a-286b-56f2-9b2c-22d5b63f416e", 00:15:27.262 "is_configured": true, 00:15:27.262 "data_offset": 2048, 00:15:27.262 "data_size": 63488 00:15:27.262 }, 00:15:27.262 { 00:15:27.262 "name": "BaseBdev2", 00:15:27.262 "uuid": "71014908-0916-5fee-85d5-2a2e4a4b4bc5", 00:15:27.262 "is_configured": true, 00:15:27.262 "data_offset": 2048, 00:15:27.262 "data_size": 63488 00:15:27.262 }, 00:15:27.262 { 00:15:27.262 "name": "BaseBdev3", 00:15:27.262 "uuid": "1cb48daf-e0ed-56fd-8811-afa342721340", 00:15:27.262 "is_configured": true, 00:15:27.262 "data_offset": 2048, 00:15:27.262 "data_size": 63488 00:15:27.262 } 00:15:27.262 ] 00:15:27.262 }' 00:15:27.262 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:27.262 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:27.521 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:27.521 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:27.521 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:27.521 [2024-11-19 15:21:17.762085] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:27.521 [2024-11-19 15:21:17.762163] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:27.521 [2024-11-19 15:21:17.762249] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:27.521 [2024-11-19 15:21:17.762320] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:27.521 [2024-11-19 15:21:17.762329] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:15:27.521 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:27.521 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:27.521 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:27.521 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:15:27.521 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:27.521 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:27.521 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:15:27.521 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:15:27.521 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:15:27.521 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:15:27.521 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:15:27.521 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:15:27.521 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:15:27.521 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:15:27.521 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:15:27.521 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:15:27.521 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:15:27.521 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:27.521 15:21:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:15:27.781 /dev/nbd0 00:15:27.781 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:15:27.781 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:15:27.781 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:15:27.781 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:15:27.781 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:15:27.781 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:15:27.781 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:15:27.781 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:15:27.781 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:15:27.781 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:15:27.781 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:27.781 1+0 records in 00:15:27.781 1+0 records out 00:15:27.781 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000448543 s, 9.1 MB/s 00:15:27.781 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:27.781 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:15:27.781 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:27.781 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:15:27.781 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:15:27.781 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:27.781 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:27.781 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:15:28.041 /dev/nbd1 00:15:28.041 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:15:28.041 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:15:28.041 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:15:28.041 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:15:28.041 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:15:28.041 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:15:28.041 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:15:28.041 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:15:28.041 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:15:28.041 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:15:28.041 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:28.041 1+0 records in 00:15:28.041 1+0 records out 00:15:28.041 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000593573 s, 6.9 MB/s 00:15:28.041 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:28.041 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:15:28.041 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:28.041 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:15:28.041 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:15:28.041 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:28.041 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:28.041 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:15:28.301 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:15:28.301 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:15:28.301 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:15:28.301 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:15:28.301 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:15:28.301 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:28.301 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:15:28.301 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:15:28.301 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:15:28.301 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:15:28.301 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:28.301 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:28.301 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:15:28.301 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:15:28.301 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:15:28.301 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:28.301 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:15:28.562 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:15:28.562 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:15:28.562 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:15:28.562 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:28.562 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:28.562 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:15:28.562 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:15:28.562 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:15:28.562 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:15:28.562 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:15:28.562 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:28.562 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:28.562 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:28.562 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:28.562 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:28.562 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:28.562 [2024-11-19 15:21:18.872877] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:28.562 [2024-11-19 15:21:18.872976] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:28.562 [2024-11-19 15:21:18.873031] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:15:28.562 [2024-11-19 15:21:18.873069] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:28.562 [2024-11-19 15:21:18.875202] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:28.562 [2024-11-19 15:21:18.875299] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:28.562 [2024-11-19 15:21:18.875393] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:15:28.562 [2024-11-19 15:21:18.875456] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:28.562 [2024-11-19 15:21:18.875599] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:28.562 [2024-11-19 15:21:18.875742] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:15:28.562 spare 00:15:28.562 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:28.562 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:15:28.562 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:28.562 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:28.822 [2024-11-19 15:21:18.975688] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:15:28.822 [2024-11-19 15:21:18.975744] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:15:28.822 [2024-11-19 15:21:18.976073] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000043d50 00:15:28.822 [2024-11-19 15:21:18.976496] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:15:28.822 [2024-11-19 15:21:18.976546] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:15:28.822 [2024-11-19 15:21:18.976740] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:28.822 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:28.822 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:28.822 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:28.822 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:28.822 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:28.822 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:28.822 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:28.822 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:28.822 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:28.822 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:28.822 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:28.822 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:28.822 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:28.822 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:28.822 15:21:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:28.822 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:28.822 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:28.822 "name": "raid_bdev1", 00:15:28.822 "uuid": "ff3a5741-c604-4f3c-98f3-f3eb4f28c032", 00:15:28.822 "strip_size_kb": 64, 00:15:28.822 "state": "online", 00:15:28.822 "raid_level": "raid5f", 00:15:28.822 "superblock": true, 00:15:28.822 "num_base_bdevs": 3, 00:15:28.822 "num_base_bdevs_discovered": 3, 00:15:28.822 "num_base_bdevs_operational": 3, 00:15:28.822 "base_bdevs_list": [ 00:15:28.822 { 00:15:28.822 "name": "spare", 00:15:28.822 "uuid": "78e4b36a-286b-56f2-9b2c-22d5b63f416e", 00:15:28.822 "is_configured": true, 00:15:28.822 "data_offset": 2048, 00:15:28.822 "data_size": 63488 00:15:28.822 }, 00:15:28.822 { 00:15:28.822 "name": "BaseBdev2", 00:15:28.822 "uuid": "71014908-0916-5fee-85d5-2a2e4a4b4bc5", 00:15:28.822 "is_configured": true, 00:15:28.822 "data_offset": 2048, 00:15:28.822 "data_size": 63488 00:15:28.822 }, 00:15:28.822 { 00:15:28.822 "name": "BaseBdev3", 00:15:28.822 "uuid": "1cb48daf-e0ed-56fd-8811-afa342721340", 00:15:28.822 "is_configured": true, 00:15:28.822 "data_offset": 2048, 00:15:28.822 "data_size": 63488 00:15:28.822 } 00:15:28.822 ] 00:15:28.822 }' 00:15:28.822 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:28.822 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:29.392 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:29.392 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:29.392 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:29.392 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:29.392 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:29.392 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:29.392 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:29.392 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:29.392 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:29.392 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:29.392 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:29.392 "name": "raid_bdev1", 00:15:29.392 "uuid": "ff3a5741-c604-4f3c-98f3-f3eb4f28c032", 00:15:29.392 "strip_size_kb": 64, 00:15:29.392 "state": "online", 00:15:29.392 "raid_level": "raid5f", 00:15:29.392 "superblock": true, 00:15:29.392 "num_base_bdevs": 3, 00:15:29.392 "num_base_bdevs_discovered": 3, 00:15:29.392 "num_base_bdevs_operational": 3, 00:15:29.392 "base_bdevs_list": [ 00:15:29.392 { 00:15:29.392 "name": "spare", 00:15:29.392 "uuid": "78e4b36a-286b-56f2-9b2c-22d5b63f416e", 00:15:29.392 "is_configured": true, 00:15:29.392 "data_offset": 2048, 00:15:29.392 "data_size": 63488 00:15:29.392 }, 00:15:29.392 { 00:15:29.392 "name": "BaseBdev2", 00:15:29.392 "uuid": "71014908-0916-5fee-85d5-2a2e4a4b4bc5", 00:15:29.392 "is_configured": true, 00:15:29.392 "data_offset": 2048, 00:15:29.392 "data_size": 63488 00:15:29.392 }, 00:15:29.392 { 00:15:29.392 "name": "BaseBdev3", 00:15:29.392 "uuid": "1cb48daf-e0ed-56fd-8811-afa342721340", 00:15:29.392 "is_configured": true, 00:15:29.392 "data_offset": 2048, 00:15:29.392 "data_size": 63488 00:15:29.392 } 00:15:29.392 ] 00:15:29.392 }' 00:15:29.392 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:29.392 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:29.392 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:29.392 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:29.392 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:29.392 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:15:29.392 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:29.392 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:29.392 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:29.392 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:15:29.392 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:15:29.392 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:29.392 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:29.392 [2024-11-19 15:21:19.631745] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:29.392 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:29.392 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:15:29.392 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:29.392 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:29.392 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:29.392 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:29.392 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:29.392 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:29.392 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:29.392 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:29.392 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:29.392 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:29.392 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:29.392 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:29.392 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:29.392 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:29.392 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:29.392 "name": "raid_bdev1", 00:15:29.392 "uuid": "ff3a5741-c604-4f3c-98f3-f3eb4f28c032", 00:15:29.392 "strip_size_kb": 64, 00:15:29.392 "state": "online", 00:15:29.392 "raid_level": "raid5f", 00:15:29.392 "superblock": true, 00:15:29.392 "num_base_bdevs": 3, 00:15:29.392 "num_base_bdevs_discovered": 2, 00:15:29.392 "num_base_bdevs_operational": 2, 00:15:29.392 "base_bdevs_list": [ 00:15:29.392 { 00:15:29.392 "name": null, 00:15:29.392 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:29.392 "is_configured": false, 00:15:29.392 "data_offset": 0, 00:15:29.392 "data_size": 63488 00:15:29.392 }, 00:15:29.392 { 00:15:29.392 "name": "BaseBdev2", 00:15:29.392 "uuid": "71014908-0916-5fee-85d5-2a2e4a4b4bc5", 00:15:29.393 "is_configured": true, 00:15:29.393 "data_offset": 2048, 00:15:29.393 "data_size": 63488 00:15:29.393 }, 00:15:29.393 { 00:15:29.393 "name": "BaseBdev3", 00:15:29.393 "uuid": "1cb48daf-e0ed-56fd-8811-afa342721340", 00:15:29.393 "is_configured": true, 00:15:29.393 "data_offset": 2048, 00:15:29.393 "data_size": 63488 00:15:29.393 } 00:15:29.393 ] 00:15:29.393 }' 00:15:29.393 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:29.393 15:21:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:29.962 15:21:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:29.962 15:21:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:29.962 15:21:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:29.962 [2024-11-19 15:21:20.138902] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:29.962 [2024-11-19 15:21:20.139104] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:15:29.962 [2024-11-19 15:21:20.139119] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:15:29.962 [2024-11-19 15:21:20.139160] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:29.962 [2024-11-19 15:21:20.143604] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000043e20 00:15:29.962 15:21:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:29.962 15:21:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:15:29.962 [2024-11-19 15:21:20.145653] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:30.901 15:21:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:30.901 15:21:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:30.901 15:21:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:30.901 15:21:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:30.901 15:21:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:30.901 15:21:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:30.901 15:21:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:30.901 15:21:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:30.901 15:21:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:30.901 15:21:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:30.901 15:21:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:30.901 "name": "raid_bdev1", 00:15:30.901 "uuid": "ff3a5741-c604-4f3c-98f3-f3eb4f28c032", 00:15:30.901 "strip_size_kb": 64, 00:15:30.901 "state": "online", 00:15:30.901 "raid_level": "raid5f", 00:15:30.901 "superblock": true, 00:15:30.901 "num_base_bdevs": 3, 00:15:30.901 "num_base_bdevs_discovered": 3, 00:15:30.901 "num_base_bdevs_operational": 3, 00:15:30.901 "process": { 00:15:30.901 "type": "rebuild", 00:15:30.901 "target": "spare", 00:15:30.901 "progress": { 00:15:30.901 "blocks": 20480, 00:15:30.901 "percent": 16 00:15:30.901 } 00:15:30.901 }, 00:15:30.901 "base_bdevs_list": [ 00:15:30.901 { 00:15:30.901 "name": "spare", 00:15:30.901 "uuid": "78e4b36a-286b-56f2-9b2c-22d5b63f416e", 00:15:30.901 "is_configured": true, 00:15:30.901 "data_offset": 2048, 00:15:30.901 "data_size": 63488 00:15:30.901 }, 00:15:30.901 { 00:15:30.901 "name": "BaseBdev2", 00:15:30.901 "uuid": "71014908-0916-5fee-85d5-2a2e4a4b4bc5", 00:15:30.901 "is_configured": true, 00:15:30.901 "data_offset": 2048, 00:15:30.901 "data_size": 63488 00:15:30.901 }, 00:15:30.901 { 00:15:30.901 "name": "BaseBdev3", 00:15:30.901 "uuid": "1cb48daf-e0ed-56fd-8811-afa342721340", 00:15:30.901 "is_configured": true, 00:15:30.901 "data_offset": 2048, 00:15:30.901 "data_size": 63488 00:15:30.901 } 00:15:30.901 ] 00:15:30.901 }' 00:15:30.901 15:21:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:31.161 15:21:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:31.161 15:21:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:31.161 15:21:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:31.161 15:21:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:15:31.161 15:21:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:31.161 15:21:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:31.161 [2024-11-19 15:21:21.309681] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:31.161 [2024-11-19 15:21:21.352143] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:31.161 [2024-11-19 15:21:21.352257] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:31.161 [2024-11-19 15:21:21.352297] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:31.161 [2024-11-19 15:21:21.352317] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:31.161 15:21:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:31.161 15:21:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:15:31.161 15:21:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:31.161 15:21:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:31.161 15:21:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:31.161 15:21:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:31.161 15:21:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:31.161 15:21:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:31.161 15:21:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:31.161 15:21:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:31.161 15:21:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:31.161 15:21:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:31.161 15:21:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:31.161 15:21:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:31.161 15:21:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:31.161 15:21:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:31.161 15:21:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:31.161 "name": "raid_bdev1", 00:15:31.161 "uuid": "ff3a5741-c604-4f3c-98f3-f3eb4f28c032", 00:15:31.161 "strip_size_kb": 64, 00:15:31.161 "state": "online", 00:15:31.161 "raid_level": "raid5f", 00:15:31.161 "superblock": true, 00:15:31.161 "num_base_bdevs": 3, 00:15:31.161 "num_base_bdevs_discovered": 2, 00:15:31.161 "num_base_bdevs_operational": 2, 00:15:31.161 "base_bdevs_list": [ 00:15:31.161 { 00:15:31.161 "name": null, 00:15:31.161 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:31.161 "is_configured": false, 00:15:31.161 "data_offset": 0, 00:15:31.161 "data_size": 63488 00:15:31.161 }, 00:15:31.161 { 00:15:31.161 "name": "BaseBdev2", 00:15:31.161 "uuid": "71014908-0916-5fee-85d5-2a2e4a4b4bc5", 00:15:31.161 "is_configured": true, 00:15:31.161 "data_offset": 2048, 00:15:31.161 "data_size": 63488 00:15:31.161 }, 00:15:31.161 { 00:15:31.161 "name": "BaseBdev3", 00:15:31.161 "uuid": "1cb48daf-e0ed-56fd-8811-afa342721340", 00:15:31.161 "is_configured": true, 00:15:31.161 "data_offset": 2048, 00:15:31.161 "data_size": 63488 00:15:31.161 } 00:15:31.161 ] 00:15:31.161 }' 00:15:31.161 15:21:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:31.161 15:21:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:31.731 15:21:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:31.731 15:21:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:31.731 15:21:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:31.731 [2024-11-19 15:21:21.828916] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:31.731 [2024-11-19 15:21:21.829044] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:31.731 [2024-11-19 15:21:21.829071] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:15:31.731 [2024-11-19 15:21:21.829080] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:31.731 [2024-11-19 15:21:21.829503] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:31.731 [2024-11-19 15:21:21.829521] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:31.731 [2024-11-19 15:21:21.829594] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:15:31.731 [2024-11-19 15:21:21.829604] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:15:31.731 [2024-11-19 15:21:21.829617] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:15:31.732 [2024-11-19 15:21:21.829643] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:31.732 spare 00:15:31.732 [2024-11-19 15:21:21.833440] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000043ef0 00:15:31.732 15:21:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:31.732 15:21:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:15:31.732 [2024-11-19 15:21:21.835524] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:32.672 15:21:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:32.672 15:21:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:32.672 15:21:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:32.672 15:21:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:32.672 15:21:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:32.672 15:21:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:32.672 15:21:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:32.672 15:21:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:32.672 15:21:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:32.672 15:21:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:32.672 15:21:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:32.672 "name": "raid_bdev1", 00:15:32.672 "uuid": "ff3a5741-c604-4f3c-98f3-f3eb4f28c032", 00:15:32.672 "strip_size_kb": 64, 00:15:32.672 "state": "online", 00:15:32.672 "raid_level": "raid5f", 00:15:32.672 "superblock": true, 00:15:32.672 "num_base_bdevs": 3, 00:15:32.672 "num_base_bdevs_discovered": 3, 00:15:32.672 "num_base_bdevs_operational": 3, 00:15:32.672 "process": { 00:15:32.672 "type": "rebuild", 00:15:32.672 "target": "spare", 00:15:32.672 "progress": { 00:15:32.672 "blocks": 20480, 00:15:32.672 "percent": 16 00:15:32.672 } 00:15:32.672 }, 00:15:32.672 "base_bdevs_list": [ 00:15:32.672 { 00:15:32.672 "name": "spare", 00:15:32.672 "uuid": "78e4b36a-286b-56f2-9b2c-22d5b63f416e", 00:15:32.672 "is_configured": true, 00:15:32.672 "data_offset": 2048, 00:15:32.672 "data_size": 63488 00:15:32.672 }, 00:15:32.672 { 00:15:32.672 "name": "BaseBdev2", 00:15:32.672 "uuid": "71014908-0916-5fee-85d5-2a2e4a4b4bc5", 00:15:32.672 "is_configured": true, 00:15:32.672 "data_offset": 2048, 00:15:32.672 "data_size": 63488 00:15:32.672 }, 00:15:32.672 { 00:15:32.672 "name": "BaseBdev3", 00:15:32.672 "uuid": "1cb48daf-e0ed-56fd-8811-afa342721340", 00:15:32.672 "is_configured": true, 00:15:32.672 "data_offset": 2048, 00:15:32.672 "data_size": 63488 00:15:32.672 } 00:15:32.672 ] 00:15:32.672 }' 00:15:32.672 15:21:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:32.672 15:21:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:32.672 15:21:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:32.672 15:21:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:32.672 15:21:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:15:32.672 15:21:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:32.672 15:21:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:32.672 [2024-11-19 15:21:22.972282] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:32.933 [2024-11-19 15:21:23.042121] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:32.933 [2024-11-19 15:21:23.042232] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:32.933 [2024-11-19 15:21:23.042268] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:32.933 [2024-11-19 15:21:23.042293] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:32.933 15:21:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:32.933 15:21:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:15:32.933 15:21:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:32.933 15:21:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:32.933 15:21:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:32.933 15:21:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:32.933 15:21:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:32.933 15:21:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:32.933 15:21:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:32.933 15:21:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:32.933 15:21:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:32.933 15:21:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:32.933 15:21:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:32.933 15:21:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:32.933 15:21:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:32.933 15:21:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:32.933 15:21:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:32.933 "name": "raid_bdev1", 00:15:32.933 "uuid": "ff3a5741-c604-4f3c-98f3-f3eb4f28c032", 00:15:32.933 "strip_size_kb": 64, 00:15:32.933 "state": "online", 00:15:32.933 "raid_level": "raid5f", 00:15:32.933 "superblock": true, 00:15:32.933 "num_base_bdevs": 3, 00:15:32.933 "num_base_bdevs_discovered": 2, 00:15:32.933 "num_base_bdevs_operational": 2, 00:15:32.933 "base_bdevs_list": [ 00:15:32.933 { 00:15:32.933 "name": null, 00:15:32.933 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:32.933 "is_configured": false, 00:15:32.933 "data_offset": 0, 00:15:32.933 "data_size": 63488 00:15:32.933 }, 00:15:32.933 { 00:15:32.933 "name": "BaseBdev2", 00:15:32.933 "uuid": "71014908-0916-5fee-85d5-2a2e4a4b4bc5", 00:15:32.933 "is_configured": true, 00:15:32.933 "data_offset": 2048, 00:15:32.933 "data_size": 63488 00:15:32.933 }, 00:15:32.933 { 00:15:32.933 "name": "BaseBdev3", 00:15:32.933 "uuid": "1cb48daf-e0ed-56fd-8811-afa342721340", 00:15:32.933 "is_configured": true, 00:15:32.933 "data_offset": 2048, 00:15:32.933 "data_size": 63488 00:15:32.933 } 00:15:32.933 ] 00:15:32.933 }' 00:15:32.933 15:21:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:32.933 15:21:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:33.193 15:21:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:33.193 15:21:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:33.193 15:21:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:33.193 15:21:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:33.193 15:21:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:33.193 15:21:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:33.193 15:21:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:33.193 15:21:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:33.193 15:21:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:33.453 15:21:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:33.453 15:21:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:33.453 "name": "raid_bdev1", 00:15:33.453 "uuid": "ff3a5741-c604-4f3c-98f3-f3eb4f28c032", 00:15:33.453 "strip_size_kb": 64, 00:15:33.453 "state": "online", 00:15:33.453 "raid_level": "raid5f", 00:15:33.453 "superblock": true, 00:15:33.453 "num_base_bdevs": 3, 00:15:33.453 "num_base_bdevs_discovered": 2, 00:15:33.453 "num_base_bdevs_operational": 2, 00:15:33.453 "base_bdevs_list": [ 00:15:33.453 { 00:15:33.453 "name": null, 00:15:33.453 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:33.453 "is_configured": false, 00:15:33.453 "data_offset": 0, 00:15:33.453 "data_size": 63488 00:15:33.453 }, 00:15:33.453 { 00:15:33.453 "name": "BaseBdev2", 00:15:33.453 "uuid": "71014908-0916-5fee-85d5-2a2e4a4b4bc5", 00:15:33.453 "is_configured": true, 00:15:33.453 "data_offset": 2048, 00:15:33.453 "data_size": 63488 00:15:33.453 }, 00:15:33.453 { 00:15:33.453 "name": "BaseBdev3", 00:15:33.453 "uuid": "1cb48daf-e0ed-56fd-8811-afa342721340", 00:15:33.453 "is_configured": true, 00:15:33.453 "data_offset": 2048, 00:15:33.453 "data_size": 63488 00:15:33.453 } 00:15:33.453 ] 00:15:33.453 }' 00:15:33.453 15:21:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:33.453 15:21:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:33.453 15:21:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:33.453 15:21:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:33.453 15:21:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:15:33.453 15:21:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:33.453 15:21:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:33.453 15:21:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:33.453 15:21:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:15:33.453 15:21:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:33.453 15:21:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:33.453 [2024-11-19 15:21:23.670706] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:15:33.453 [2024-11-19 15:21:23.670818] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:33.453 [2024-11-19 15:21:23.670861] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:15:33.453 [2024-11-19 15:21:23.670897] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:33.453 [2024-11-19 15:21:23.671298] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:33.453 [2024-11-19 15:21:23.671357] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:15:33.453 [2024-11-19 15:21:23.671451] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:15:33.454 [2024-11-19 15:21:23.671493] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:15:33.454 [2024-11-19 15:21:23.671539] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:15:33.454 [2024-11-19 15:21:23.671568] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:15:33.454 BaseBdev1 00:15:33.454 15:21:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:33.454 15:21:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:15:34.394 15:21:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:15:34.394 15:21:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:34.394 15:21:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:34.394 15:21:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:34.394 15:21:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:34.394 15:21:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:34.394 15:21:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:34.394 15:21:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:34.394 15:21:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:34.394 15:21:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:34.394 15:21:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:34.394 15:21:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:34.394 15:21:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:34.394 15:21:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:34.394 15:21:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:34.654 15:21:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:34.654 "name": "raid_bdev1", 00:15:34.654 "uuid": "ff3a5741-c604-4f3c-98f3-f3eb4f28c032", 00:15:34.654 "strip_size_kb": 64, 00:15:34.654 "state": "online", 00:15:34.654 "raid_level": "raid5f", 00:15:34.654 "superblock": true, 00:15:34.654 "num_base_bdevs": 3, 00:15:34.654 "num_base_bdevs_discovered": 2, 00:15:34.654 "num_base_bdevs_operational": 2, 00:15:34.654 "base_bdevs_list": [ 00:15:34.654 { 00:15:34.654 "name": null, 00:15:34.654 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:34.654 "is_configured": false, 00:15:34.654 "data_offset": 0, 00:15:34.654 "data_size": 63488 00:15:34.654 }, 00:15:34.654 { 00:15:34.654 "name": "BaseBdev2", 00:15:34.654 "uuid": "71014908-0916-5fee-85d5-2a2e4a4b4bc5", 00:15:34.654 "is_configured": true, 00:15:34.654 "data_offset": 2048, 00:15:34.654 "data_size": 63488 00:15:34.654 }, 00:15:34.654 { 00:15:34.654 "name": "BaseBdev3", 00:15:34.654 "uuid": "1cb48daf-e0ed-56fd-8811-afa342721340", 00:15:34.654 "is_configured": true, 00:15:34.654 "data_offset": 2048, 00:15:34.654 "data_size": 63488 00:15:34.654 } 00:15:34.654 ] 00:15:34.654 }' 00:15:34.654 15:21:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:34.654 15:21:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:34.915 15:21:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:34.915 15:21:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:34.915 15:21:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:34.915 15:21:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:34.915 15:21:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:34.915 15:21:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:34.915 15:21:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:34.915 15:21:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:34.915 15:21:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:34.915 15:21:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:34.915 15:21:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:34.915 "name": "raid_bdev1", 00:15:34.915 "uuid": "ff3a5741-c604-4f3c-98f3-f3eb4f28c032", 00:15:34.915 "strip_size_kb": 64, 00:15:34.915 "state": "online", 00:15:34.915 "raid_level": "raid5f", 00:15:34.915 "superblock": true, 00:15:34.915 "num_base_bdevs": 3, 00:15:34.915 "num_base_bdevs_discovered": 2, 00:15:34.915 "num_base_bdevs_operational": 2, 00:15:34.915 "base_bdevs_list": [ 00:15:34.915 { 00:15:34.915 "name": null, 00:15:34.915 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:34.915 "is_configured": false, 00:15:34.915 "data_offset": 0, 00:15:34.915 "data_size": 63488 00:15:34.915 }, 00:15:34.915 { 00:15:34.915 "name": "BaseBdev2", 00:15:34.915 "uuid": "71014908-0916-5fee-85d5-2a2e4a4b4bc5", 00:15:34.915 "is_configured": true, 00:15:34.915 "data_offset": 2048, 00:15:34.915 "data_size": 63488 00:15:34.915 }, 00:15:34.915 { 00:15:34.915 "name": "BaseBdev3", 00:15:34.915 "uuid": "1cb48daf-e0ed-56fd-8811-afa342721340", 00:15:34.915 "is_configured": true, 00:15:34.915 "data_offset": 2048, 00:15:34.915 "data_size": 63488 00:15:34.915 } 00:15:34.915 ] 00:15:34.915 }' 00:15:34.915 15:21:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:34.915 15:21:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:35.175 15:21:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:35.175 15:21:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:35.175 15:21:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:35.175 15:21:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@652 -- # local es=0 00:15:35.175 15:21:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:35.175 15:21:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:15:35.175 15:21:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:15:35.175 15:21:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:15:35.175 15:21:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:15:35.175 15:21:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:35.175 15:21:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:35.175 15:21:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:35.175 [2024-11-19 15:21:25.296020] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:35.175 [2024-11-19 15:21:25.296173] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:15:35.175 [2024-11-19 15:21:25.296185] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:15:35.175 request: 00:15:35.175 { 00:15:35.175 "base_bdev": "BaseBdev1", 00:15:35.175 "raid_bdev": "raid_bdev1", 00:15:35.175 "method": "bdev_raid_add_base_bdev", 00:15:35.175 "req_id": 1 00:15:35.175 } 00:15:35.175 Got JSON-RPC error response 00:15:35.175 response: 00:15:35.175 { 00:15:35.175 "code": -22, 00:15:35.175 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:15:35.175 } 00:15:35.175 15:21:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:15:35.175 15:21:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@655 -- # es=1 00:15:35.175 15:21:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:15:35.175 15:21:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:15:35.175 15:21:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:15:35.175 15:21:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:15:36.115 15:21:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:15:36.115 15:21:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:36.115 15:21:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:36.115 15:21:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:36.115 15:21:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:36.115 15:21:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:36.115 15:21:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:36.115 15:21:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:36.115 15:21:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:36.115 15:21:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:36.115 15:21:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:36.115 15:21:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:36.115 15:21:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:36.115 15:21:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:36.115 15:21:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:36.115 15:21:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:36.115 "name": "raid_bdev1", 00:15:36.115 "uuid": "ff3a5741-c604-4f3c-98f3-f3eb4f28c032", 00:15:36.115 "strip_size_kb": 64, 00:15:36.115 "state": "online", 00:15:36.115 "raid_level": "raid5f", 00:15:36.115 "superblock": true, 00:15:36.115 "num_base_bdevs": 3, 00:15:36.115 "num_base_bdevs_discovered": 2, 00:15:36.115 "num_base_bdevs_operational": 2, 00:15:36.115 "base_bdevs_list": [ 00:15:36.115 { 00:15:36.115 "name": null, 00:15:36.115 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:36.115 "is_configured": false, 00:15:36.115 "data_offset": 0, 00:15:36.115 "data_size": 63488 00:15:36.115 }, 00:15:36.115 { 00:15:36.115 "name": "BaseBdev2", 00:15:36.115 "uuid": "71014908-0916-5fee-85d5-2a2e4a4b4bc5", 00:15:36.115 "is_configured": true, 00:15:36.115 "data_offset": 2048, 00:15:36.115 "data_size": 63488 00:15:36.115 }, 00:15:36.115 { 00:15:36.115 "name": "BaseBdev3", 00:15:36.115 "uuid": "1cb48daf-e0ed-56fd-8811-afa342721340", 00:15:36.115 "is_configured": true, 00:15:36.115 "data_offset": 2048, 00:15:36.115 "data_size": 63488 00:15:36.115 } 00:15:36.115 ] 00:15:36.115 }' 00:15:36.115 15:21:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:36.115 15:21:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:36.685 15:21:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:36.685 15:21:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:36.685 15:21:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:36.685 15:21:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:36.685 15:21:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:36.685 15:21:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:36.685 15:21:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:36.685 15:21:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:36.685 15:21:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:36.685 15:21:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:36.685 15:21:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:36.685 "name": "raid_bdev1", 00:15:36.685 "uuid": "ff3a5741-c604-4f3c-98f3-f3eb4f28c032", 00:15:36.685 "strip_size_kb": 64, 00:15:36.685 "state": "online", 00:15:36.685 "raid_level": "raid5f", 00:15:36.685 "superblock": true, 00:15:36.685 "num_base_bdevs": 3, 00:15:36.685 "num_base_bdevs_discovered": 2, 00:15:36.685 "num_base_bdevs_operational": 2, 00:15:36.685 "base_bdevs_list": [ 00:15:36.685 { 00:15:36.685 "name": null, 00:15:36.685 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:36.685 "is_configured": false, 00:15:36.685 "data_offset": 0, 00:15:36.685 "data_size": 63488 00:15:36.685 }, 00:15:36.685 { 00:15:36.685 "name": "BaseBdev2", 00:15:36.685 "uuid": "71014908-0916-5fee-85d5-2a2e4a4b4bc5", 00:15:36.685 "is_configured": true, 00:15:36.685 "data_offset": 2048, 00:15:36.685 "data_size": 63488 00:15:36.685 }, 00:15:36.685 { 00:15:36.685 "name": "BaseBdev3", 00:15:36.685 "uuid": "1cb48daf-e0ed-56fd-8811-afa342721340", 00:15:36.685 "is_configured": true, 00:15:36.685 "data_offset": 2048, 00:15:36.685 "data_size": 63488 00:15:36.685 } 00:15:36.685 ] 00:15:36.685 }' 00:15:36.685 15:21:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:36.685 15:21:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:36.685 15:21:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:36.685 15:21:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:36.685 15:21:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 92522 00:15:36.685 15:21:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@954 -- # '[' -z 92522 ']' 00:15:36.685 15:21:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@958 -- # kill -0 92522 00:15:36.685 15:21:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@959 -- # uname 00:15:36.685 15:21:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:15:36.685 15:21:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 92522 00:15:36.685 15:21:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:15:36.685 killing process with pid 92522 00:15:36.685 Received shutdown signal, test time was about 60.000000 seconds 00:15:36.685 00:15:36.685 Latency(us) 00:15:36.685 [2024-11-19T15:21:27.024Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:15:36.685 [2024-11-19T15:21:27.024Z] =================================================================================================================== 00:15:36.685 [2024-11-19T15:21:27.024Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:15:36.685 15:21:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:15:36.685 15:21:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 92522' 00:15:36.685 15:21:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@973 -- # kill 92522 00:15:36.685 [2024-11-19 15:21:26.928881] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:36.685 [2024-11-19 15:21:26.929008] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:36.685 [2024-11-19 15:21:26.929073] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:36.685 [2024-11-19 15:21:26.929082] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:15:36.685 15:21:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@978 -- # wait 92522 00:15:36.685 [2024-11-19 15:21:26.970787] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:36.946 15:21:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:15:36.947 ************************************ 00:15:36.947 END TEST raid5f_rebuild_test_sb 00:15:36.947 ************************************ 00:15:36.947 00:15:36.947 real 0m21.646s 00:15:36.947 user 0m28.232s 00:15:36.947 sys 0m2.768s 00:15:36.947 15:21:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:15:36.947 15:21:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:36.947 15:21:27 bdev_raid -- bdev/bdev_raid.sh@985 -- # for n in {3..4} 00:15:36.947 15:21:27 bdev_raid -- bdev/bdev_raid.sh@986 -- # run_test raid5f_state_function_test raid_state_function_test raid5f 4 false 00:15:36.947 15:21:27 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:15:36.947 15:21:27 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:15:36.947 15:21:27 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:36.947 ************************************ 00:15:36.947 START TEST raid5f_state_function_test 00:15:36.947 ************************************ 00:15:36.947 15:21:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid5f 4 false 00:15:36.947 15:21:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:15:36.947 15:21:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:15:36.947 15:21:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:15:36.947 15:21:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:15:36.947 15:21:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:15:36.947 15:21:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:36.947 15:21:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:15:36.947 15:21:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:15:36.947 15:21:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:36.947 15:21:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:15:36.947 15:21:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:15:36.947 15:21:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:36.947 15:21:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:15:36.947 15:21:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:15:36.947 15:21:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:36.947 15:21:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:15:36.947 15:21:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:15:36.947 15:21:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:36.947 15:21:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:15:36.947 15:21:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:15:36.947 15:21:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:15:36.947 15:21:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:15:36.947 15:21:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:15:36.947 15:21:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:15:36.947 15:21:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:15:36.947 15:21:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:15:36.947 15:21:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:15:36.947 Process raid pid: 93252 00:15:36.947 15:21:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:15:36.947 15:21:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:15:36.947 15:21:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=93252 00:15:36.947 15:21:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 93252' 00:15:36.947 15:21:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 93252 00:15:36.947 15:21:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:15:36.947 15:21:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 93252 ']' 00:15:36.947 15:21:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:36.947 15:21:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:15:36.947 15:21:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:36.947 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:36.947 15:21:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:15:36.947 15:21:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:37.207 [2024-11-19 15:21:27.356985] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:15:37.208 [2024-11-19 15:21:27.357211] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:15:37.208 [2024-11-19 15:21:27.498367] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:37.208 [2024-11-19 15:21:27.524008] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:37.468 [2024-11-19 15:21:27.567496] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:37.468 [2024-11-19 15:21:27.567530] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:38.038 15:21:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:15:38.038 15:21:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:15:38.038 15:21:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:15:38.038 15:21:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:38.038 15:21:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:38.038 [2024-11-19 15:21:28.169245] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:15:38.038 [2024-11-19 15:21:28.169353] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:15:38.038 [2024-11-19 15:21:28.169384] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:38.038 [2024-11-19 15:21:28.169408] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:38.038 [2024-11-19 15:21:28.169427] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:15:38.038 [2024-11-19 15:21:28.169449] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:15:38.038 [2024-11-19 15:21:28.169466] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:15:38.038 [2024-11-19 15:21:28.169511] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:15:38.038 15:21:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:38.038 15:21:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:38.038 15:21:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:38.038 15:21:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:38.038 15:21:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:38.038 15:21:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:38.038 15:21:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:38.038 15:21:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:38.038 15:21:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:38.038 15:21:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:38.038 15:21:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:38.038 15:21:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:38.038 15:21:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:38.038 15:21:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:38.038 15:21:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:38.038 15:21:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:38.038 15:21:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:38.038 "name": "Existed_Raid", 00:15:38.038 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:38.038 "strip_size_kb": 64, 00:15:38.038 "state": "configuring", 00:15:38.038 "raid_level": "raid5f", 00:15:38.038 "superblock": false, 00:15:38.038 "num_base_bdevs": 4, 00:15:38.038 "num_base_bdevs_discovered": 0, 00:15:38.038 "num_base_bdevs_operational": 4, 00:15:38.038 "base_bdevs_list": [ 00:15:38.038 { 00:15:38.038 "name": "BaseBdev1", 00:15:38.038 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:38.038 "is_configured": false, 00:15:38.038 "data_offset": 0, 00:15:38.038 "data_size": 0 00:15:38.038 }, 00:15:38.038 { 00:15:38.038 "name": "BaseBdev2", 00:15:38.038 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:38.039 "is_configured": false, 00:15:38.039 "data_offset": 0, 00:15:38.039 "data_size": 0 00:15:38.039 }, 00:15:38.039 { 00:15:38.039 "name": "BaseBdev3", 00:15:38.039 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:38.039 "is_configured": false, 00:15:38.039 "data_offset": 0, 00:15:38.039 "data_size": 0 00:15:38.039 }, 00:15:38.039 { 00:15:38.039 "name": "BaseBdev4", 00:15:38.039 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:38.039 "is_configured": false, 00:15:38.039 "data_offset": 0, 00:15:38.039 "data_size": 0 00:15:38.039 } 00:15:38.039 ] 00:15:38.039 }' 00:15:38.039 15:21:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:38.039 15:21:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:38.299 15:21:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:15:38.299 15:21:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:38.299 15:21:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:38.299 [2024-11-19 15:21:28.624358] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:15:38.299 [2024-11-19 15:21:28.624435] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:15:38.299 15:21:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:38.299 15:21:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:15:38.299 15:21:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:38.299 15:21:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:38.299 [2024-11-19 15:21:28.632373] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:15:38.299 [2024-11-19 15:21:28.632449] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:15:38.299 [2024-11-19 15:21:28.632479] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:38.299 [2024-11-19 15:21:28.632501] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:38.299 [2024-11-19 15:21:28.632533] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:15:38.299 [2024-11-19 15:21:28.632554] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:15:38.299 [2024-11-19 15:21:28.632574] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:15:38.299 [2024-11-19 15:21:28.632585] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:15:38.299 15:21:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:38.299 15:21:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:15:38.299 15:21:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:38.299 15:21:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:38.559 BaseBdev1 00:15:38.559 [2024-11-19 15:21:28.649472] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:38.560 15:21:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:38.560 15:21:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:15:38.560 15:21:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:15:38.560 15:21:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:38.560 15:21:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:15:38.560 15:21:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:38.560 15:21:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:38.560 15:21:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:38.560 15:21:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:38.560 15:21:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:38.560 15:21:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:38.560 15:21:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:15:38.560 15:21:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:38.560 15:21:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:38.560 [ 00:15:38.560 { 00:15:38.560 "name": "BaseBdev1", 00:15:38.560 "aliases": [ 00:15:38.560 "69080764-2d7e-490a-9c01-d35d4e48459f" 00:15:38.560 ], 00:15:38.560 "product_name": "Malloc disk", 00:15:38.560 "block_size": 512, 00:15:38.560 "num_blocks": 65536, 00:15:38.560 "uuid": "69080764-2d7e-490a-9c01-d35d4e48459f", 00:15:38.560 "assigned_rate_limits": { 00:15:38.560 "rw_ios_per_sec": 0, 00:15:38.560 "rw_mbytes_per_sec": 0, 00:15:38.560 "r_mbytes_per_sec": 0, 00:15:38.560 "w_mbytes_per_sec": 0 00:15:38.560 }, 00:15:38.560 "claimed": true, 00:15:38.560 "claim_type": "exclusive_write", 00:15:38.560 "zoned": false, 00:15:38.560 "supported_io_types": { 00:15:38.560 "read": true, 00:15:38.560 "write": true, 00:15:38.560 "unmap": true, 00:15:38.560 "flush": true, 00:15:38.560 "reset": true, 00:15:38.560 "nvme_admin": false, 00:15:38.560 "nvme_io": false, 00:15:38.560 "nvme_io_md": false, 00:15:38.560 "write_zeroes": true, 00:15:38.560 "zcopy": true, 00:15:38.560 "get_zone_info": false, 00:15:38.560 "zone_management": false, 00:15:38.560 "zone_append": false, 00:15:38.560 "compare": false, 00:15:38.560 "compare_and_write": false, 00:15:38.560 "abort": true, 00:15:38.560 "seek_hole": false, 00:15:38.560 "seek_data": false, 00:15:38.560 "copy": true, 00:15:38.560 "nvme_iov_md": false 00:15:38.560 }, 00:15:38.560 "memory_domains": [ 00:15:38.560 { 00:15:38.560 "dma_device_id": "system", 00:15:38.560 "dma_device_type": 1 00:15:38.560 }, 00:15:38.560 { 00:15:38.560 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:38.560 "dma_device_type": 2 00:15:38.560 } 00:15:38.560 ], 00:15:38.560 "driver_specific": {} 00:15:38.560 } 00:15:38.560 ] 00:15:38.560 15:21:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:38.560 15:21:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:15:38.560 15:21:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:38.560 15:21:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:38.560 15:21:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:38.560 15:21:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:38.560 15:21:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:38.560 15:21:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:38.560 15:21:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:38.560 15:21:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:38.560 15:21:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:38.560 15:21:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:38.560 15:21:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:38.560 15:21:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:38.560 15:21:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:38.560 15:21:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:38.560 15:21:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:38.560 15:21:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:38.560 "name": "Existed_Raid", 00:15:38.560 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:38.560 "strip_size_kb": 64, 00:15:38.560 "state": "configuring", 00:15:38.560 "raid_level": "raid5f", 00:15:38.560 "superblock": false, 00:15:38.560 "num_base_bdevs": 4, 00:15:38.560 "num_base_bdevs_discovered": 1, 00:15:38.560 "num_base_bdevs_operational": 4, 00:15:38.560 "base_bdevs_list": [ 00:15:38.560 { 00:15:38.560 "name": "BaseBdev1", 00:15:38.560 "uuid": "69080764-2d7e-490a-9c01-d35d4e48459f", 00:15:38.560 "is_configured": true, 00:15:38.560 "data_offset": 0, 00:15:38.560 "data_size": 65536 00:15:38.560 }, 00:15:38.560 { 00:15:38.560 "name": "BaseBdev2", 00:15:38.560 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:38.560 "is_configured": false, 00:15:38.560 "data_offset": 0, 00:15:38.560 "data_size": 0 00:15:38.560 }, 00:15:38.560 { 00:15:38.560 "name": "BaseBdev3", 00:15:38.560 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:38.560 "is_configured": false, 00:15:38.560 "data_offset": 0, 00:15:38.560 "data_size": 0 00:15:38.560 }, 00:15:38.560 { 00:15:38.560 "name": "BaseBdev4", 00:15:38.560 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:38.560 "is_configured": false, 00:15:38.560 "data_offset": 0, 00:15:38.560 "data_size": 0 00:15:38.560 } 00:15:38.560 ] 00:15:38.560 }' 00:15:38.560 15:21:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:38.560 15:21:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:38.820 15:21:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:15:38.820 15:21:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:38.820 15:21:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:38.820 [2024-11-19 15:21:29.120676] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:15:38.820 [2024-11-19 15:21:29.120757] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:15:38.820 15:21:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:38.820 15:21:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:15:38.821 15:21:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:38.821 15:21:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:38.821 [2024-11-19 15:21:29.132695] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:38.821 [2024-11-19 15:21:29.134562] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:38.821 [2024-11-19 15:21:29.134600] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:38.821 [2024-11-19 15:21:29.134609] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:15:38.821 [2024-11-19 15:21:29.134617] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:15:38.821 [2024-11-19 15:21:29.134623] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:15:38.821 [2024-11-19 15:21:29.134631] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:15:38.821 15:21:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:38.821 15:21:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:15:38.821 15:21:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:15:38.821 15:21:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:38.821 15:21:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:38.821 15:21:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:38.821 15:21:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:38.821 15:21:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:38.821 15:21:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:38.821 15:21:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:38.821 15:21:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:38.821 15:21:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:38.821 15:21:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:38.821 15:21:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:38.821 15:21:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:38.821 15:21:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:38.821 15:21:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:39.091 15:21:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:39.091 15:21:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:39.091 "name": "Existed_Raid", 00:15:39.091 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:39.091 "strip_size_kb": 64, 00:15:39.091 "state": "configuring", 00:15:39.091 "raid_level": "raid5f", 00:15:39.091 "superblock": false, 00:15:39.091 "num_base_bdevs": 4, 00:15:39.091 "num_base_bdevs_discovered": 1, 00:15:39.091 "num_base_bdevs_operational": 4, 00:15:39.091 "base_bdevs_list": [ 00:15:39.091 { 00:15:39.091 "name": "BaseBdev1", 00:15:39.091 "uuid": "69080764-2d7e-490a-9c01-d35d4e48459f", 00:15:39.091 "is_configured": true, 00:15:39.091 "data_offset": 0, 00:15:39.091 "data_size": 65536 00:15:39.091 }, 00:15:39.091 { 00:15:39.091 "name": "BaseBdev2", 00:15:39.091 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:39.091 "is_configured": false, 00:15:39.091 "data_offset": 0, 00:15:39.091 "data_size": 0 00:15:39.091 }, 00:15:39.091 { 00:15:39.091 "name": "BaseBdev3", 00:15:39.092 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:39.092 "is_configured": false, 00:15:39.092 "data_offset": 0, 00:15:39.092 "data_size": 0 00:15:39.092 }, 00:15:39.092 { 00:15:39.092 "name": "BaseBdev4", 00:15:39.092 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:39.092 "is_configured": false, 00:15:39.092 "data_offset": 0, 00:15:39.092 "data_size": 0 00:15:39.092 } 00:15:39.092 ] 00:15:39.092 }' 00:15:39.092 15:21:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:39.092 15:21:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:39.355 15:21:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:15:39.355 15:21:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:39.355 15:21:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:39.355 [2024-11-19 15:21:29.587051] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:39.355 BaseBdev2 00:15:39.355 15:21:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:39.355 15:21:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:15:39.355 15:21:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:15:39.355 15:21:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:39.355 15:21:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:15:39.355 15:21:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:39.355 15:21:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:39.355 15:21:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:39.355 15:21:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:39.355 15:21:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:39.355 15:21:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:39.355 15:21:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:15:39.355 15:21:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:39.355 15:21:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:39.355 [ 00:15:39.355 { 00:15:39.355 "name": "BaseBdev2", 00:15:39.355 "aliases": [ 00:15:39.355 "4684bc74-4820-44de-bc00-23a7c69fb2c3" 00:15:39.355 ], 00:15:39.355 "product_name": "Malloc disk", 00:15:39.355 "block_size": 512, 00:15:39.355 "num_blocks": 65536, 00:15:39.355 "uuid": "4684bc74-4820-44de-bc00-23a7c69fb2c3", 00:15:39.355 "assigned_rate_limits": { 00:15:39.355 "rw_ios_per_sec": 0, 00:15:39.355 "rw_mbytes_per_sec": 0, 00:15:39.355 "r_mbytes_per_sec": 0, 00:15:39.355 "w_mbytes_per_sec": 0 00:15:39.355 }, 00:15:39.355 "claimed": true, 00:15:39.355 "claim_type": "exclusive_write", 00:15:39.355 "zoned": false, 00:15:39.355 "supported_io_types": { 00:15:39.355 "read": true, 00:15:39.355 "write": true, 00:15:39.355 "unmap": true, 00:15:39.355 "flush": true, 00:15:39.355 "reset": true, 00:15:39.355 "nvme_admin": false, 00:15:39.355 "nvme_io": false, 00:15:39.355 "nvme_io_md": false, 00:15:39.355 "write_zeroes": true, 00:15:39.355 "zcopy": true, 00:15:39.355 "get_zone_info": false, 00:15:39.355 "zone_management": false, 00:15:39.355 "zone_append": false, 00:15:39.355 "compare": false, 00:15:39.355 "compare_and_write": false, 00:15:39.355 "abort": true, 00:15:39.355 "seek_hole": false, 00:15:39.355 "seek_data": false, 00:15:39.355 "copy": true, 00:15:39.355 "nvme_iov_md": false 00:15:39.355 }, 00:15:39.355 "memory_domains": [ 00:15:39.355 { 00:15:39.355 "dma_device_id": "system", 00:15:39.355 "dma_device_type": 1 00:15:39.355 }, 00:15:39.355 { 00:15:39.355 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:39.355 "dma_device_type": 2 00:15:39.355 } 00:15:39.355 ], 00:15:39.355 "driver_specific": {} 00:15:39.355 } 00:15:39.355 ] 00:15:39.355 15:21:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:39.355 15:21:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:15:39.355 15:21:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:15:39.355 15:21:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:15:39.355 15:21:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:39.355 15:21:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:39.355 15:21:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:39.355 15:21:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:39.355 15:21:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:39.355 15:21:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:39.355 15:21:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:39.355 15:21:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:39.355 15:21:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:39.355 15:21:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:39.355 15:21:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:39.355 15:21:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:39.355 15:21:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:39.355 15:21:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:39.355 15:21:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:39.355 15:21:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:39.355 "name": "Existed_Raid", 00:15:39.355 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:39.355 "strip_size_kb": 64, 00:15:39.355 "state": "configuring", 00:15:39.355 "raid_level": "raid5f", 00:15:39.355 "superblock": false, 00:15:39.355 "num_base_bdevs": 4, 00:15:39.355 "num_base_bdevs_discovered": 2, 00:15:39.355 "num_base_bdevs_operational": 4, 00:15:39.355 "base_bdevs_list": [ 00:15:39.355 { 00:15:39.355 "name": "BaseBdev1", 00:15:39.355 "uuid": "69080764-2d7e-490a-9c01-d35d4e48459f", 00:15:39.355 "is_configured": true, 00:15:39.355 "data_offset": 0, 00:15:39.355 "data_size": 65536 00:15:39.355 }, 00:15:39.355 { 00:15:39.355 "name": "BaseBdev2", 00:15:39.355 "uuid": "4684bc74-4820-44de-bc00-23a7c69fb2c3", 00:15:39.355 "is_configured": true, 00:15:39.355 "data_offset": 0, 00:15:39.355 "data_size": 65536 00:15:39.355 }, 00:15:39.355 { 00:15:39.355 "name": "BaseBdev3", 00:15:39.355 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:39.355 "is_configured": false, 00:15:39.355 "data_offset": 0, 00:15:39.355 "data_size": 0 00:15:39.355 }, 00:15:39.355 { 00:15:39.355 "name": "BaseBdev4", 00:15:39.355 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:39.355 "is_configured": false, 00:15:39.355 "data_offset": 0, 00:15:39.355 "data_size": 0 00:15:39.355 } 00:15:39.355 ] 00:15:39.355 }' 00:15:39.355 15:21:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:39.355 15:21:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:39.925 15:21:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:15:39.925 15:21:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:39.925 15:21:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:39.925 [2024-11-19 15:21:30.099727] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:15:39.925 BaseBdev3 00:15:39.925 15:21:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:39.925 15:21:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:15:39.925 15:21:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:15:39.925 15:21:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:39.925 15:21:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:15:39.925 15:21:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:39.925 15:21:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:39.926 15:21:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:39.926 15:21:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:39.926 15:21:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:39.926 15:21:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:39.926 15:21:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:15:39.926 15:21:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:39.926 15:21:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:39.926 [ 00:15:39.926 { 00:15:39.926 "name": "BaseBdev3", 00:15:39.926 "aliases": [ 00:15:39.926 "01a11f1e-f1a9-41cd-9c7b-53cbc3d70348" 00:15:39.926 ], 00:15:39.926 "product_name": "Malloc disk", 00:15:39.926 "block_size": 512, 00:15:39.926 "num_blocks": 65536, 00:15:39.926 "uuid": "01a11f1e-f1a9-41cd-9c7b-53cbc3d70348", 00:15:39.926 "assigned_rate_limits": { 00:15:39.926 "rw_ios_per_sec": 0, 00:15:39.926 "rw_mbytes_per_sec": 0, 00:15:39.926 "r_mbytes_per_sec": 0, 00:15:39.926 "w_mbytes_per_sec": 0 00:15:39.926 }, 00:15:39.926 "claimed": true, 00:15:39.926 "claim_type": "exclusive_write", 00:15:39.926 "zoned": false, 00:15:39.926 "supported_io_types": { 00:15:39.926 "read": true, 00:15:39.926 "write": true, 00:15:39.926 "unmap": true, 00:15:39.926 "flush": true, 00:15:39.926 "reset": true, 00:15:39.926 "nvme_admin": false, 00:15:39.926 "nvme_io": false, 00:15:39.926 "nvme_io_md": false, 00:15:39.926 "write_zeroes": true, 00:15:39.926 "zcopy": true, 00:15:39.926 "get_zone_info": false, 00:15:39.926 "zone_management": false, 00:15:39.926 "zone_append": false, 00:15:39.926 "compare": false, 00:15:39.926 "compare_and_write": false, 00:15:39.926 "abort": true, 00:15:39.926 "seek_hole": false, 00:15:39.926 "seek_data": false, 00:15:39.926 "copy": true, 00:15:39.926 "nvme_iov_md": false 00:15:39.926 }, 00:15:39.926 "memory_domains": [ 00:15:39.926 { 00:15:39.926 "dma_device_id": "system", 00:15:39.926 "dma_device_type": 1 00:15:39.926 }, 00:15:39.926 { 00:15:39.926 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:39.926 "dma_device_type": 2 00:15:39.926 } 00:15:39.926 ], 00:15:39.926 "driver_specific": {} 00:15:39.926 } 00:15:39.926 ] 00:15:39.926 15:21:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:39.926 15:21:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:15:39.926 15:21:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:15:39.926 15:21:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:15:39.926 15:21:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:39.926 15:21:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:39.926 15:21:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:39.926 15:21:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:39.926 15:21:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:39.926 15:21:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:39.926 15:21:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:39.926 15:21:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:39.926 15:21:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:39.926 15:21:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:39.926 15:21:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:39.926 15:21:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:39.926 15:21:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:39.926 15:21:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:39.926 15:21:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:39.926 15:21:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:39.926 "name": "Existed_Raid", 00:15:39.926 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:39.926 "strip_size_kb": 64, 00:15:39.926 "state": "configuring", 00:15:39.926 "raid_level": "raid5f", 00:15:39.926 "superblock": false, 00:15:39.926 "num_base_bdevs": 4, 00:15:39.926 "num_base_bdevs_discovered": 3, 00:15:39.926 "num_base_bdevs_operational": 4, 00:15:39.926 "base_bdevs_list": [ 00:15:39.926 { 00:15:39.926 "name": "BaseBdev1", 00:15:39.926 "uuid": "69080764-2d7e-490a-9c01-d35d4e48459f", 00:15:39.926 "is_configured": true, 00:15:39.926 "data_offset": 0, 00:15:39.926 "data_size": 65536 00:15:39.926 }, 00:15:39.926 { 00:15:39.926 "name": "BaseBdev2", 00:15:39.926 "uuid": "4684bc74-4820-44de-bc00-23a7c69fb2c3", 00:15:39.926 "is_configured": true, 00:15:39.926 "data_offset": 0, 00:15:39.926 "data_size": 65536 00:15:39.926 }, 00:15:39.926 { 00:15:39.926 "name": "BaseBdev3", 00:15:39.926 "uuid": "01a11f1e-f1a9-41cd-9c7b-53cbc3d70348", 00:15:39.926 "is_configured": true, 00:15:39.926 "data_offset": 0, 00:15:39.926 "data_size": 65536 00:15:39.926 }, 00:15:39.926 { 00:15:39.926 "name": "BaseBdev4", 00:15:39.926 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:39.926 "is_configured": false, 00:15:39.926 "data_offset": 0, 00:15:39.926 "data_size": 0 00:15:39.926 } 00:15:39.926 ] 00:15:39.926 }' 00:15:39.926 15:21:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:39.926 15:21:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:40.497 15:21:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:15:40.497 15:21:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:40.497 15:21:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:40.497 [2024-11-19 15:21:30.597996] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:15:40.497 [2024-11-19 15:21:30.598117] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:15:40.497 [2024-11-19 15:21:30.598130] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:15:40.497 [2024-11-19 15:21:30.598414] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:15:40.497 [2024-11-19 15:21:30.598876] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:15:40.497 [2024-11-19 15:21:30.598889] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:15:40.497 [2024-11-19 15:21:30.599120] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:40.497 BaseBdev4 00:15:40.497 15:21:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:40.497 15:21:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:15:40.497 15:21:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:15:40.497 15:21:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:40.497 15:21:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:15:40.497 15:21:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:40.497 15:21:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:40.497 15:21:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:40.497 15:21:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:40.497 15:21:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:40.497 15:21:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:40.497 15:21:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:15:40.497 15:21:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:40.497 15:21:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:40.497 [ 00:15:40.497 { 00:15:40.497 "name": "BaseBdev4", 00:15:40.497 "aliases": [ 00:15:40.497 "6600b475-ae13-46ee-95a7-5ef6a38eec6a" 00:15:40.497 ], 00:15:40.497 "product_name": "Malloc disk", 00:15:40.497 "block_size": 512, 00:15:40.497 "num_blocks": 65536, 00:15:40.497 "uuid": "6600b475-ae13-46ee-95a7-5ef6a38eec6a", 00:15:40.497 "assigned_rate_limits": { 00:15:40.497 "rw_ios_per_sec": 0, 00:15:40.497 "rw_mbytes_per_sec": 0, 00:15:40.497 "r_mbytes_per_sec": 0, 00:15:40.497 "w_mbytes_per_sec": 0 00:15:40.497 }, 00:15:40.497 "claimed": true, 00:15:40.497 "claim_type": "exclusive_write", 00:15:40.497 "zoned": false, 00:15:40.497 "supported_io_types": { 00:15:40.497 "read": true, 00:15:40.497 "write": true, 00:15:40.497 "unmap": true, 00:15:40.497 "flush": true, 00:15:40.497 "reset": true, 00:15:40.497 "nvme_admin": false, 00:15:40.497 "nvme_io": false, 00:15:40.497 "nvme_io_md": false, 00:15:40.497 "write_zeroes": true, 00:15:40.497 "zcopy": true, 00:15:40.497 "get_zone_info": false, 00:15:40.497 "zone_management": false, 00:15:40.497 "zone_append": false, 00:15:40.497 "compare": false, 00:15:40.497 "compare_and_write": false, 00:15:40.497 "abort": true, 00:15:40.497 "seek_hole": false, 00:15:40.497 "seek_data": false, 00:15:40.497 "copy": true, 00:15:40.497 "nvme_iov_md": false 00:15:40.497 }, 00:15:40.497 "memory_domains": [ 00:15:40.497 { 00:15:40.497 "dma_device_id": "system", 00:15:40.497 "dma_device_type": 1 00:15:40.497 }, 00:15:40.497 { 00:15:40.497 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:40.497 "dma_device_type": 2 00:15:40.497 } 00:15:40.497 ], 00:15:40.497 "driver_specific": {} 00:15:40.497 } 00:15:40.497 ] 00:15:40.497 15:21:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:40.497 15:21:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:15:40.497 15:21:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:15:40.497 15:21:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:15:40.497 15:21:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:15:40.497 15:21:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:40.497 15:21:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:40.497 15:21:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:40.497 15:21:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:40.497 15:21:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:40.497 15:21:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:40.497 15:21:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:40.497 15:21:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:40.497 15:21:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:40.497 15:21:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:40.497 15:21:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:40.497 15:21:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:40.497 15:21:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:40.497 15:21:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:40.497 15:21:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:40.497 "name": "Existed_Raid", 00:15:40.497 "uuid": "7d6ce7fe-fc08-48cc-a006-e9d259656134", 00:15:40.497 "strip_size_kb": 64, 00:15:40.497 "state": "online", 00:15:40.497 "raid_level": "raid5f", 00:15:40.497 "superblock": false, 00:15:40.497 "num_base_bdevs": 4, 00:15:40.497 "num_base_bdevs_discovered": 4, 00:15:40.497 "num_base_bdevs_operational": 4, 00:15:40.497 "base_bdevs_list": [ 00:15:40.497 { 00:15:40.497 "name": "BaseBdev1", 00:15:40.497 "uuid": "69080764-2d7e-490a-9c01-d35d4e48459f", 00:15:40.497 "is_configured": true, 00:15:40.497 "data_offset": 0, 00:15:40.497 "data_size": 65536 00:15:40.497 }, 00:15:40.497 { 00:15:40.497 "name": "BaseBdev2", 00:15:40.497 "uuid": "4684bc74-4820-44de-bc00-23a7c69fb2c3", 00:15:40.497 "is_configured": true, 00:15:40.497 "data_offset": 0, 00:15:40.497 "data_size": 65536 00:15:40.497 }, 00:15:40.497 { 00:15:40.498 "name": "BaseBdev3", 00:15:40.498 "uuid": "01a11f1e-f1a9-41cd-9c7b-53cbc3d70348", 00:15:40.498 "is_configured": true, 00:15:40.498 "data_offset": 0, 00:15:40.498 "data_size": 65536 00:15:40.498 }, 00:15:40.498 { 00:15:40.498 "name": "BaseBdev4", 00:15:40.498 "uuid": "6600b475-ae13-46ee-95a7-5ef6a38eec6a", 00:15:40.498 "is_configured": true, 00:15:40.498 "data_offset": 0, 00:15:40.498 "data_size": 65536 00:15:40.498 } 00:15:40.498 ] 00:15:40.498 }' 00:15:40.498 15:21:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:40.498 15:21:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:40.757 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:15:40.757 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:15:40.757 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:40.757 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:40.757 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:15:40.757 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:40.757 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:15:40.757 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:40.757 15:21:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:40.757 15:21:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:40.757 [2024-11-19 15:21:31.081369] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:41.123 15:21:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.123 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:41.123 "name": "Existed_Raid", 00:15:41.123 "aliases": [ 00:15:41.123 "7d6ce7fe-fc08-48cc-a006-e9d259656134" 00:15:41.123 ], 00:15:41.123 "product_name": "Raid Volume", 00:15:41.123 "block_size": 512, 00:15:41.124 "num_blocks": 196608, 00:15:41.124 "uuid": "7d6ce7fe-fc08-48cc-a006-e9d259656134", 00:15:41.124 "assigned_rate_limits": { 00:15:41.124 "rw_ios_per_sec": 0, 00:15:41.124 "rw_mbytes_per_sec": 0, 00:15:41.124 "r_mbytes_per_sec": 0, 00:15:41.124 "w_mbytes_per_sec": 0 00:15:41.124 }, 00:15:41.124 "claimed": false, 00:15:41.124 "zoned": false, 00:15:41.124 "supported_io_types": { 00:15:41.124 "read": true, 00:15:41.124 "write": true, 00:15:41.124 "unmap": false, 00:15:41.124 "flush": false, 00:15:41.124 "reset": true, 00:15:41.124 "nvme_admin": false, 00:15:41.124 "nvme_io": false, 00:15:41.124 "nvme_io_md": false, 00:15:41.124 "write_zeroes": true, 00:15:41.124 "zcopy": false, 00:15:41.124 "get_zone_info": false, 00:15:41.124 "zone_management": false, 00:15:41.124 "zone_append": false, 00:15:41.124 "compare": false, 00:15:41.124 "compare_and_write": false, 00:15:41.124 "abort": false, 00:15:41.124 "seek_hole": false, 00:15:41.124 "seek_data": false, 00:15:41.124 "copy": false, 00:15:41.124 "nvme_iov_md": false 00:15:41.124 }, 00:15:41.124 "driver_specific": { 00:15:41.124 "raid": { 00:15:41.124 "uuid": "7d6ce7fe-fc08-48cc-a006-e9d259656134", 00:15:41.124 "strip_size_kb": 64, 00:15:41.124 "state": "online", 00:15:41.124 "raid_level": "raid5f", 00:15:41.124 "superblock": false, 00:15:41.124 "num_base_bdevs": 4, 00:15:41.124 "num_base_bdevs_discovered": 4, 00:15:41.124 "num_base_bdevs_operational": 4, 00:15:41.124 "base_bdevs_list": [ 00:15:41.124 { 00:15:41.124 "name": "BaseBdev1", 00:15:41.124 "uuid": "69080764-2d7e-490a-9c01-d35d4e48459f", 00:15:41.124 "is_configured": true, 00:15:41.124 "data_offset": 0, 00:15:41.124 "data_size": 65536 00:15:41.124 }, 00:15:41.124 { 00:15:41.124 "name": "BaseBdev2", 00:15:41.124 "uuid": "4684bc74-4820-44de-bc00-23a7c69fb2c3", 00:15:41.124 "is_configured": true, 00:15:41.124 "data_offset": 0, 00:15:41.124 "data_size": 65536 00:15:41.124 }, 00:15:41.124 { 00:15:41.124 "name": "BaseBdev3", 00:15:41.124 "uuid": "01a11f1e-f1a9-41cd-9c7b-53cbc3d70348", 00:15:41.124 "is_configured": true, 00:15:41.124 "data_offset": 0, 00:15:41.124 "data_size": 65536 00:15:41.124 }, 00:15:41.124 { 00:15:41.124 "name": "BaseBdev4", 00:15:41.124 "uuid": "6600b475-ae13-46ee-95a7-5ef6a38eec6a", 00:15:41.124 "is_configured": true, 00:15:41.124 "data_offset": 0, 00:15:41.124 "data_size": 65536 00:15:41.124 } 00:15:41.124 ] 00:15:41.124 } 00:15:41.124 } 00:15:41.124 }' 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:15:41.124 BaseBdev2 00:15:41.124 BaseBdev3 00:15:41.124 BaseBdev4' 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:41.124 [2024-11-19 15:21:31.392688] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.124 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:41.124 "name": "Existed_Raid", 00:15:41.124 "uuid": "7d6ce7fe-fc08-48cc-a006-e9d259656134", 00:15:41.124 "strip_size_kb": 64, 00:15:41.124 "state": "online", 00:15:41.124 "raid_level": "raid5f", 00:15:41.124 "superblock": false, 00:15:41.124 "num_base_bdevs": 4, 00:15:41.124 "num_base_bdevs_discovered": 3, 00:15:41.124 "num_base_bdevs_operational": 3, 00:15:41.124 "base_bdevs_list": [ 00:15:41.124 { 00:15:41.124 "name": null, 00:15:41.124 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:41.124 "is_configured": false, 00:15:41.124 "data_offset": 0, 00:15:41.124 "data_size": 65536 00:15:41.124 }, 00:15:41.124 { 00:15:41.124 "name": "BaseBdev2", 00:15:41.124 "uuid": "4684bc74-4820-44de-bc00-23a7c69fb2c3", 00:15:41.124 "is_configured": true, 00:15:41.124 "data_offset": 0, 00:15:41.124 "data_size": 65536 00:15:41.124 }, 00:15:41.124 { 00:15:41.125 "name": "BaseBdev3", 00:15:41.125 "uuid": "01a11f1e-f1a9-41cd-9c7b-53cbc3d70348", 00:15:41.125 "is_configured": true, 00:15:41.125 "data_offset": 0, 00:15:41.125 "data_size": 65536 00:15:41.125 }, 00:15:41.125 { 00:15:41.125 "name": "BaseBdev4", 00:15:41.125 "uuid": "6600b475-ae13-46ee-95a7-5ef6a38eec6a", 00:15:41.125 "is_configured": true, 00:15:41.125 "data_offset": 0, 00:15:41.125 "data_size": 65536 00:15:41.125 } 00:15:41.125 ] 00:15:41.125 }' 00:15:41.125 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:41.403 15:21:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:41.662 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:15:41.662 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:15:41.662 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:41.662 15:21:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.662 15:21:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:41.662 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:15:41.662 15:21:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.662 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:15:41.662 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:15:41.662 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:15:41.662 15:21:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.662 15:21:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:41.662 [2024-11-19 15:21:31.887326] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:15:41.662 [2024-11-19 15:21:31.887476] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:41.662 [2024-11-19 15:21:31.898713] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:41.662 15:21:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.662 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:15:41.662 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:15:41.662 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:15:41.662 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:41.662 15:21:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.662 15:21:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:41.662 15:21:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.662 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:15:41.662 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:15:41.662 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:15:41.662 15:21:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.662 15:21:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:41.663 [2024-11-19 15:21:31.942666] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:15:41.663 15:21:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.663 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:15:41.663 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:15:41.663 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:41.663 15:21:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:15:41.663 15:21:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.663 15:21:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:41.663 15:21:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.923 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:15:41.923 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:15:41.923 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:15:41.923 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.923 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:41.923 [2024-11-19 15:21:32.013806] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:15:41.923 [2024-11-19 15:21:32.013903] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:15:41.923 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.923 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:15:41.923 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:15:41.923 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:41.923 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:15:41.923 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.923 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:41.923 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.923 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:15:41.923 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:41.924 BaseBdev2 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:41.924 [ 00:15:41.924 { 00:15:41.924 "name": "BaseBdev2", 00:15:41.924 "aliases": [ 00:15:41.924 "03a46437-56c7-47ec-82ff-d54f5d087971" 00:15:41.924 ], 00:15:41.924 "product_name": "Malloc disk", 00:15:41.924 "block_size": 512, 00:15:41.924 "num_blocks": 65536, 00:15:41.924 "uuid": "03a46437-56c7-47ec-82ff-d54f5d087971", 00:15:41.924 "assigned_rate_limits": { 00:15:41.924 "rw_ios_per_sec": 0, 00:15:41.924 "rw_mbytes_per_sec": 0, 00:15:41.924 "r_mbytes_per_sec": 0, 00:15:41.924 "w_mbytes_per_sec": 0 00:15:41.924 }, 00:15:41.924 "claimed": false, 00:15:41.924 "zoned": false, 00:15:41.924 "supported_io_types": { 00:15:41.924 "read": true, 00:15:41.924 "write": true, 00:15:41.924 "unmap": true, 00:15:41.924 "flush": true, 00:15:41.924 "reset": true, 00:15:41.924 "nvme_admin": false, 00:15:41.924 "nvme_io": false, 00:15:41.924 "nvme_io_md": false, 00:15:41.924 "write_zeroes": true, 00:15:41.924 "zcopy": true, 00:15:41.924 "get_zone_info": false, 00:15:41.924 "zone_management": false, 00:15:41.924 "zone_append": false, 00:15:41.924 "compare": false, 00:15:41.924 "compare_and_write": false, 00:15:41.924 "abort": true, 00:15:41.924 "seek_hole": false, 00:15:41.924 "seek_data": false, 00:15:41.924 "copy": true, 00:15:41.924 "nvme_iov_md": false 00:15:41.924 }, 00:15:41.924 "memory_domains": [ 00:15:41.924 { 00:15:41.924 "dma_device_id": "system", 00:15:41.924 "dma_device_type": 1 00:15:41.924 }, 00:15:41.924 { 00:15:41.924 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:41.924 "dma_device_type": 2 00:15:41.924 } 00:15:41.924 ], 00:15:41.924 "driver_specific": {} 00:15:41.924 } 00:15:41.924 ] 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:41.924 BaseBdev3 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:41.924 [ 00:15:41.924 { 00:15:41.924 "name": "BaseBdev3", 00:15:41.924 "aliases": [ 00:15:41.924 "f846b0df-70a7-44e6-9dfa-d0d03d606d82" 00:15:41.924 ], 00:15:41.924 "product_name": "Malloc disk", 00:15:41.924 "block_size": 512, 00:15:41.924 "num_blocks": 65536, 00:15:41.924 "uuid": "f846b0df-70a7-44e6-9dfa-d0d03d606d82", 00:15:41.924 "assigned_rate_limits": { 00:15:41.924 "rw_ios_per_sec": 0, 00:15:41.924 "rw_mbytes_per_sec": 0, 00:15:41.924 "r_mbytes_per_sec": 0, 00:15:41.924 "w_mbytes_per_sec": 0 00:15:41.924 }, 00:15:41.924 "claimed": false, 00:15:41.924 "zoned": false, 00:15:41.924 "supported_io_types": { 00:15:41.924 "read": true, 00:15:41.924 "write": true, 00:15:41.924 "unmap": true, 00:15:41.924 "flush": true, 00:15:41.924 "reset": true, 00:15:41.924 "nvme_admin": false, 00:15:41.924 "nvme_io": false, 00:15:41.924 "nvme_io_md": false, 00:15:41.924 "write_zeroes": true, 00:15:41.924 "zcopy": true, 00:15:41.924 "get_zone_info": false, 00:15:41.924 "zone_management": false, 00:15:41.924 "zone_append": false, 00:15:41.924 "compare": false, 00:15:41.924 "compare_and_write": false, 00:15:41.924 "abort": true, 00:15:41.924 "seek_hole": false, 00:15:41.924 "seek_data": false, 00:15:41.924 "copy": true, 00:15:41.924 "nvme_iov_md": false 00:15:41.924 }, 00:15:41.924 "memory_domains": [ 00:15:41.924 { 00:15:41.924 "dma_device_id": "system", 00:15:41.924 "dma_device_type": 1 00:15:41.924 }, 00:15:41.924 { 00:15:41.924 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:41.924 "dma_device_type": 2 00:15:41.924 } 00:15:41.924 ], 00:15:41.924 "driver_specific": {} 00:15:41.924 } 00:15:41.924 ] 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:41.924 BaseBdev4 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.924 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:41.924 [ 00:15:41.924 { 00:15:41.924 "name": "BaseBdev4", 00:15:41.924 "aliases": [ 00:15:41.924 "a72dbccf-b414-4e92-9f37-af8f4174426f" 00:15:41.924 ], 00:15:41.924 "product_name": "Malloc disk", 00:15:41.924 "block_size": 512, 00:15:41.924 "num_blocks": 65536, 00:15:41.924 "uuid": "a72dbccf-b414-4e92-9f37-af8f4174426f", 00:15:41.924 "assigned_rate_limits": { 00:15:41.924 "rw_ios_per_sec": 0, 00:15:41.924 "rw_mbytes_per_sec": 0, 00:15:41.924 "r_mbytes_per_sec": 0, 00:15:41.924 "w_mbytes_per_sec": 0 00:15:41.924 }, 00:15:41.924 "claimed": false, 00:15:41.924 "zoned": false, 00:15:41.924 "supported_io_types": { 00:15:41.924 "read": true, 00:15:41.925 "write": true, 00:15:41.925 "unmap": true, 00:15:41.925 "flush": true, 00:15:41.925 "reset": true, 00:15:41.925 "nvme_admin": false, 00:15:41.925 "nvme_io": false, 00:15:41.925 "nvme_io_md": false, 00:15:41.925 "write_zeroes": true, 00:15:41.925 "zcopy": true, 00:15:41.925 "get_zone_info": false, 00:15:41.925 "zone_management": false, 00:15:41.925 "zone_append": false, 00:15:41.925 "compare": false, 00:15:41.925 "compare_and_write": false, 00:15:41.925 "abort": true, 00:15:41.925 "seek_hole": false, 00:15:41.925 "seek_data": false, 00:15:41.925 "copy": true, 00:15:41.925 "nvme_iov_md": false 00:15:41.925 }, 00:15:41.925 "memory_domains": [ 00:15:41.925 { 00:15:41.925 "dma_device_id": "system", 00:15:41.925 "dma_device_type": 1 00:15:41.925 }, 00:15:41.925 { 00:15:41.925 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:41.925 "dma_device_type": 2 00:15:41.925 } 00:15:41.925 ], 00:15:41.925 "driver_specific": {} 00:15:41.925 } 00:15:41.925 ] 00:15:41.925 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.925 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:15:41.925 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:15:41.925 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:15:41.925 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:15:41.925 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.925 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:41.925 [2024-11-19 15:21:32.244872] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:15:41.925 [2024-11-19 15:21:32.244956] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:15:41.925 [2024-11-19 15:21:32.245012] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:41.925 [2024-11-19 15:21:32.246781] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:15:41.925 [2024-11-19 15:21:32.246879] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:15:41.925 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.925 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:41.925 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:41.925 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:41.925 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:41.925 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:41.925 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:41.925 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:41.925 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:41.925 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:41.925 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:41.925 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:41.925 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:41.925 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.925 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:42.185 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:42.185 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:42.185 "name": "Existed_Raid", 00:15:42.185 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:42.185 "strip_size_kb": 64, 00:15:42.185 "state": "configuring", 00:15:42.185 "raid_level": "raid5f", 00:15:42.185 "superblock": false, 00:15:42.185 "num_base_bdevs": 4, 00:15:42.185 "num_base_bdevs_discovered": 3, 00:15:42.185 "num_base_bdevs_operational": 4, 00:15:42.185 "base_bdevs_list": [ 00:15:42.185 { 00:15:42.185 "name": "BaseBdev1", 00:15:42.185 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:42.185 "is_configured": false, 00:15:42.185 "data_offset": 0, 00:15:42.185 "data_size": 0 00:15:42.185 }, 00:15:42.185 { 00:15:42.185 "name": "BaseBdev2", 00:15:42.185 "uuid": "03a46437-56c7-47ec-82ff-d54f5d087971", 00:15:42.185 "is_configured": true, 00:15:42.185 "data_offset": 0, 00:15:42.185 "data_size": 65536 00:15:42.185 }, 00:15:42.185 { 00:15:42.185 "name": "BaseBdev3", 00:15:42.185 "uuid": "f846b0df-70a7-44e6-9dfa-d0d03d606d82", 00:15:42.185 "is_configured": true, 00:15:42.185 "data_offset": 0, 00:15:42.185 "data_size": 65536 00:15:42.185 }, 00:15:42.185 { 00:15:42.185 "name": "BaseBdev4", 00:15:42.185 "uuid": "a72dbccf-b414-4e92-9f37-af8f4174426f", 00:15:42.185 "is_configured": true, 00:15:42.185 "data_offset": 0, 00:15:42.185 "data_size": 65536 00:15:42.185 } 00:15:42.185 ] 00:15:42.185 }' 00:15:42.185 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:42.185 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:42.445 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:15:42.445 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:42.445 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:42.445 [2024-11-19 15:21:32.728041] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:15:42.445 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:42.445 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:42.445 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:42.445 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:42.445 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:42.445 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:42.445 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:42.445 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:42.445 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:42.445 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:42.445 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:42.445 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:42.445 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:42.445 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:42.445 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:42.445 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:42.445 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:42.445 "name": "Existed_Raid", 00:15:42.445 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:42.445 "strip_size_kb": 64, 00:15:42.445 "state": "configuring", 00:15:42.445 "raid_level": "raid5f", 00:15:42.445 "superblock": false, 00:15:42.445 "num_base_bdevs": 4, 00:15:42.445 "num_base_bdevs_discovered": 2, 00:15:42.445 "num_base_bdevs_operational": 4, 00:15:42.445 "base_bdevs_list": [ 00:15:42.445 { 00:15:42.445 "name": "BaseBdev1", 00:15:42.445 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:42.445 "is_configured": false, 00:15:42.445 "data_offset": 0, 00:15:42.445 "data_size": 0 00:15:42.445 }, 00:15:42.445 { 00:15:42.445 "name": null, 00:15:42.445 "uuid": "03a46437-56c7-47ec-82ff-d54f5d087971", 00:15:42.445 "is_configured": false, 00:15:42.445 "data_offset": 0, 00:15:42.445 "data_size": 65536 00:15:42.445 }, 00:15:42.445 { 00:15:42.445 "name": "BaseBdev3", 00:15:42.446 "uuid": "f846b0df-70a7-44e6-9dfa-d0d03d606d82", 00:15:42.446 "is_configured": true, 00:15:42.446 "data_offset": 0, 00:15:42.446 "data_size": 65536 00:15:42.446 }, 00:15:42.446 { 00:15:42.446 "name": "BaseBdev4", 00:15:42.446 "uuid": "a72dbccf-b414-4e92-9f37-af8f4174426f", 00:15:42.446 "is_configured": true, 00:15:42.446 "data_offset": 0, 00:15:42.446 "data_size": 65536 00:15:42.446 } 00:15:42.446 ] 00:15:42.446 }' 00:15:42.705 15:21:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:42.705 15:21:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:42.965 15:21:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:42.965 15:21:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:15:42.965 15:21:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:42.965 15:21:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:42.965 15:21:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:42.965 15:21:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:15:42.965 15:21:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:15:42.965 15:21:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:42.965 15:21:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:42.965 [2024-11-19 15:21:33.250239] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:42.965 BaseBdev1 00:15:42.965 15:21:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:42.965 15:21:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:15:42.965 15:21:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:15:42.965 15:21:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:42.965 15:21:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:15:42.965 15:21:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:42.965 15:21:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:42.965 15:21:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:42.965 15:21:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:42.965 15:21:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:42.965 15:21:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:42.965 15:21:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:15:42.965 15:21:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:42.965 15:21:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:42.965 [ 00:15:42.965 { 00:15:42.965 "name": "BaseBdev1", 00:15:42.965 "aliases": [ 00:15:42.965 "bf21cf7a-ddbe-4728-9752-6bf57f2e0c1f" 00:15:42.965 ], 00:15:42.965 "product_name": "Malloc disk", 00:15:42.965 "block_size": 512, 00:15:42.965 "num_blocks": 65536, 00:15:42.965 "uuid": "bf21cf7a-ddbe-4728-9752-6bf57f2e0c1f", 00:15:42.965 "assigned_rate_limits": { 00:15:42.965 "rw_ios_per_sec": 0, 00:15:42.965 "rw_mbytes_per_sec": 0, 00:15:42.965 "r_mbytes_per_sec": 0, 00:15:42.965 "w_mbytes_per_sec": 0 00:15:42.965 }, 00:15:42.965 "claimed": true, 00:15:42.965 "claim_type": "exclusive_write", 00:15:42.965 "zoned": false, 00:15:42.965 "supported_io_types": { 00:15:42.965 "read": true, 00:15:42.965 "write": true, 00:15:42.965 "unmap": true, 00:15:42.965 "flush": true, 00:15:42.965 "reset": true, 00:15:42.965 "nvme_admin": false, 00:15:42.965 "nvme_io": false, 00:15:42.965 "nvme_io_md": false, 00:15:42.965 "write_zeroes": true, 00:15:42.965 "zcopy": true, 00:15:42.965 "get_zone_info": false, 00:15:42.965 "zone_management": false, 00:15:42.965 "zone_append": false, 00:15:42.965 "compare": false, 00:15:42.965 "compare_and_write": false, 00:15:42.965 "abort": true, 00:15:42.965 "seek_hole": false, 00:15:42.965 "seek_data": false, 00:15:42.965 "copy": true, 00:15:42.965 "nvme_iov_md": false 00:15:42.965 }, 00:15:42.966 "memory_domains": [ 00:15:42.966 { 00:15:42.966 "dma_device_id": "system", 00:15:42.966 "dma_device_type": 1 00:15:42.966 }, 00:15:42.966 { 00:15:42.966 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:42.966 "dma_device_type": 2 00:15:42.966 } 00:15:42.966 ], 00:15:42.966 "driver_specific": {} 00:15:42.966 } 00:15:42.966 ] 00:15:42.966 15:21:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:42.966 15:21:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:15:42.966 15:21:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:42.966 15:21:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:42.966 15:21:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:42.966 15:21:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:42.966 15:21:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:42.966 15:21:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:42.966 15:21:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:42.966 15:21:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:42.966 15:21:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:42.966 15:21:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:42.966 15:21:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:42.966 15:21:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:42.966 15:21:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:42.966 15:21:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:43.226 15:21:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:43.226 15:21:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:43.226 "name": "Existed_Raid", 00:15:43.226 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:43.226 "strip_size_kb": 64, 00:15:43.226 "state": "configuring", 00:15:43.226 "raid_level": "raid5f", 00:15:43.226 "superblock": false, 00:15:43.226 "num_base_bdevs": 4, 00:15:43.226 "num_base_bdevs_discovered": 3, 00:15:43.226 "num_base_bdevs_operational": 4, 00:15:43.226 "base_bdevs_list": [ 00:15:43.226 { 00:15:43.226 "name": "BaseBdev1", 00:15:43.226 "uuid": "bf21cf7a-ddbe-4728-9752-6bf57f2e0c1f", 00:15:43.226 "is_configured": true, 00:15:43.226 "data_offset": 0, 00:15:43.226 "data_size": 65536 00:15:43.226 }, 00:15:43.226 { 00:15:43.226 "name": null, 00:15:43.226 "uuid": "03a46437-56c7-47ec-82ff-d54f5d087971", 00:15:43.226 "is_configured": false, 00:15:43.226 "data_offset": 0, 00:15:43.226 "data_size": 65536 00:15:43.226 }, 00:15:43.226 { 00:15:43.226 "name": "BaseBdev3", 00:15:43.226 "uuid": "f846b0df-70a7-44e6-9dfa-d0d03d606d82", 00:15:43.226 "is_configured": true, 00:15:43.226 "data_offset": 0, 00:15:43.226 "data_size": 65536 00:15:43.226 }, 00:15:43.226 { 00:15:43.226 "name": "BaseBdev4", 00:15:43.226 "uuid": "a72dbccf-b414-4e92-9f37-af8f4174426f", 00:15:43.226 "is_configured": true, 00:15:43.226 "data_offset": 0, 00:15:43.226 "data_size": 65536 00:15:43.226 } 00:15:43.226 ] 00:15:43.226 }' 00:15:43.226 15:21:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:43.226 15:21:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:43.486 15:21:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:15:43.486 15:21:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:43.486 15:21:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:43.486 15:21:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:43.486 15:21:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:43.486 15:21:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:15:43.486 15:21:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:15:43.486 15:21:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:43.486 15:21:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:43.486 [2024-11-19 15:21:33.793357] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:15:43.486 15:21:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:43.486 15:21:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:43.486 15:21:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:43.486 15:21:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:43.486 15:21:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:43.486 15:21:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:43.486 15:21:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:43.486 15:21:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:43.486 15:21:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:43.486 15:21:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:43.486 15:21:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:43.486 15:21:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:43.486 15:21:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:43.486 15:21:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:43.486 15:21:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:43.746 15:21:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:43.746 15:21:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:43.746 "name": "Existed_Raid", 00:15:43.746 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:43.746 "strip_size_kb": 64, 00:15:43.746 "state": "configuring", 00:15:43.746 "raid_level": "raid5f", 00:15:43.746 "superblock": false, 00:15:43.746 "num_base_bdevs": 4, 00:15:43.746 "num_base_bdevs_discovered": 2, 00:15:43.746 "num_base_bdevs_operational": 4, 00:15:43.746 "base_bdevs_list": [ 00:15:43.746 { 00:15:43.746 "name": "BaseBdev1", 00:15:43.746 "uuid": "bf21cf7a-ddbe-4728-9752-6bf57f2e0c1f", 00:15:43.746 "is_configured": true, 00:15:43.746 "data_offset": 0, 00:15:43.746 "data_size": 65536 00:15:43.746 }, 00:15:43.746 { 00:15:43.746 "name": null, 00:15:43.746 "uuid": "03a46437-56c7-47ec-82ff-d54f5d087971", 00:15:43.746 "is_configured": false, 00:15:43.746 "data_offset": 0, 00:15:43.746 "data_size": 65536 00:15:43.746 }, 00:15:43.746 { 00:15:43.746 "name": null, 00:15:43.746 "uuid": "f846b0df-70a7-44e6-9dfa-d0d03d606d82", 00:15:43.746 "is_configured": false, 00:15:43.746 "data_offset": 0, 00:15:43.746 "data_size": 65536 00:15:43.746 }, 00:15:43.746 { 00:15:43.746 "name": "BaseBdev4", 00:15:43.746 "uuid": "a72dbccf-b414-4e92-9f37-af8f4174426f", 00:15:43.746 "is_configured": true, 00:15:43.746 "data_offset": 0, 00:15:43.746 "data_size": 65536 00:15:43.746 } 00:15:43.746 ] 00:15:43.746 }' 00:15:43.746 15:21:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:43.746 15:21:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:44.006 15:21:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:44.006 15:21:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.006 15:21:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:44.006 15:21:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:15:44.006 15:21:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.006 15:21:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:15:44.006 15:21:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:15:44.006 15:21:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.006 15:21:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:44.006 [2024-11-19 15:21:34.320456] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:15:44.006 15:21:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.006 15:21:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:44.006 15:21:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:44.006 15:21:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:44.006 15:21:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:44.006 15:21:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:44.006 15:21:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:44.006 15:21:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:44.006 15:21:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:44.006 15:21:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:44.006 15:21:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:44.006 15:21:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:44.006 15:21:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.006 15:21:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:44.006 15:21:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:44.267 15:21:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.267 15:21:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:44.267 "name": "Existed_Raid", 00:15:44.267 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:44.267 "strip_size_kb": 64, 00:15:44.267 "state": "configuring", 00:15:44.267 "raid_level": "raid5f", 00:15:44.267 "superblock": false, 00:15:44.267 "num_base_bdevs": 4, 00:15:44.267 "num_base_bdevs_discovered": 3, 00:15:44.267 "num_base_bdevs_operational": 4, 00:15:44.267 "base_bdevs_list": [ 00:15:44.267 { 00:15:44.267 "name": "BaseBdev1", 00:15:44.267 "uuid": "bf21cf7a-ddbe-4728-9752-6bf57f2e0c1f", 00:15:44.267 "is_configured": true, 00:15:44.267 "data_offset": 0, 00:15:44.267 "data_size": 65536 00:15:44.267 }, 00:15:44.267 { 00:15:44.267 "name": null, 00:15:44.267 "uuid": "03a46437-56c7-47ec-82ff-d54f5d087971", 00:15:44.267 "is_configured": false, 00:15:44.267 "data_offset": 0, 00:15:44.267 "data_size": 65536 00:15:44.267 }, 00:15:44.267 { 00:15:44.267 "name": "BaseBdev3", 00:15:44.267 "uuid": "f846b0df-70a7-44e6-9dfa-d0d03d606d82", 00:15:44.267 "is_configured": true, 00:15:44.267 "data_offset": 0, 00:15:44.267 "data_size": 65536 00:15:44.267 }, 00:15:44.267 { 00:15:44.267 "name": "BaseBdev4", 00:15:44.267 "uuid": "a72dbccf-b414-4e92-9f37-af8f4174426f", 00:15:44.267 "is_configured": true, 00:15:44.267 "data_offset": 0, 00:15:44.267 "data_size": 65536 00:15:44.267 } 00:15:44.267 ] 00:15:44.267 }' 00:15:44.267 15:21:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:44.267 15:21:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:44.527 15:21:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:44.527 15:21:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.527 15:21:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:44.527 15:21:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:15:44.527 15:21:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.527 15:21:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:15:44.527 15:21:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:15:44.527 15:21:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.527 15:21:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:44.527 [2024-11-19 15:21:34.851770] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:44.787 15:21:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.788 15:21:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:44.788 15:21:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:44.788 15:21:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:44.788 15:21:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:44.788 15:21:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:44.788 15:21:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:44.788 15:21:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:44.788 15:21:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:44.788 15:21:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:44.788 15:21:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:44.788 15:21:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:44.788 15:21:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:44.788 15:21:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.788 15:21:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:44.788 15:21:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.788 15:21:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:44.788 "name": "Existed_Raid", 00:15:44.788 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:44.788 "strip_size_kb": 64, 00:15:44.788 "state": "configuring", 00:15:44.788 "raid_level": "raid5f", 00:15:44.788 "superblock": false, 00:15:44.788 "num_base_bdevs": 4, 00:15:44.788 "num_base_bdevs_discovered": 2, 00:15:44.788 "num_base_bdevs_operational": 4, 00:15:44.788 "base_bdevs_list": [ 00:15:44.788 { 00:15:44.788 "name": null, 00:15:44.788 "uuid": "bf21cf7a-ddbe-4728-9752-6bf57f2e0c1f", 00:15:44.788 "is_configured": false, 00:15:44.788 "data_offset": 0, 00:15:44.788 "data_size": 65536 00:15:44.788 }, 00:15:44.788 { 00:15:44.788 "name": null, 00:15:44.788 "uuid": "03a46437-56c7-47ec-82ff-d54f5d087971", 00:15:44.788 "is_configured": false, 00:15:44.788 "data_offset": 0, 00:15:44.788 "data_size": 65536 00:15:44.788 }, 00:15:44.788 { 00:15:44.788 "name": "BaseBdev3", 00:15:44.788 "uuid": "f846b0df-70a7-44e6-9dfa-d0d03d606d82", 00:15:44.788 "is_configured": true, 00:15:44.788 "data_offset": 0, 00:15:44.788 "data_size": 65536 00:15:44.788 }, 00:15:44.788 { 00:15:44.788 "name": "BaseBdev4", 00:15:44.788 "uuid": "a72dbccf-b414-4e92-9f37-af8f4174426f", 00:15:44.788 "is_configured": true, 00:15:44.788 "data_offset": 0, 00:15:44.788 "data_size": 65536 00:15:44.788 } 00:15:44.788 ] 00:15:44.788 }' 00:15:44.788 15:21:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:44.788 15:21:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:45.048 15:21:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:45.048 15:21:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:15:45.048 15:21:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:45.048 15:21:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:45.048 15:21:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:45.048 15:21:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:15:45.048 15:21:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:15:45.048 15:21:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:45.048 15:21:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:45.048 [2024-11-19 15:21:35.365516] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:45.048 15:21:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:45.048 15:21:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:45.048 15:21:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:45.048 15:21:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:45.048 15:21:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:45.048 15:21:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:45.048 15:21:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:45.048 15:21:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:45.048 15:21:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:45.048 15:21:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:45.048 15:21:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:45.048 15:21:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:45.048 15:21:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:45.048 15:21:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:45.048 15:21:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:45.308 15:21:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:45.308 15:21:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:45.308 "name": "Existed_Raid", 00:15:45.308 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:45.308 "strip_size_kb": 64, 00:15:45.308 "state": "configuring", 00:15:45.308 "raid_level": "raid5f", 00:15:45.308 "superblock": false, 00:15:45.308 "num_base_bdevs": 4, 00:15:45.308 "num_base_bdevs_discovered": 3, 00:15:45.308 "num_base_bdevs_operational": 4, 00:15:45.308 "base_bdevs_list": [ 00:15:45.308 { 00:15:45.308 "name": null, 00:15:45.308 "uuid": "bf21cf7a-ddbe-4728-9752-6bf57f2e0c1f", 00:15:45.308 "is_configured": false, 00:15:45.308 "data_offset": 0, 00:15:45.308 "data_size": 65536 00:15:45.308 }, 00:15:45.308 { 00:15:45.308 "name": "BaseBdev2", 00:15:45.308 "uuid": "03a46437-56c7-47ec-82ff-d54f5d087971", 00:15:45.308 "is_configured": true, 00:15:45.308 "data_offset": 0, 00:15:45.308 "data_size": 65536 00:15:45.308 }, 00:15:45.308 { 00:15:45.308 "name": "BaseBdev3", 00:15:45.308 "uuid": "f846b0df-70a7-44e6-9dfa-d0d03d606d82", 00:15:45.308 "is_configured": true, 00:15:45.308 "data_offset": 0, 00:15:45.308 "data_size": 65536 00:15:45.308 }, 00:15:45.308 { 00:15:45.308 "name": "BaseBdev4", 00:15:45.308 "uuid": "a72dbccf-b414-4e92-9f37-af8f4174426f", 00:15:45.308 "is_configured": true, 00:15:45.308 "data_offset": 0, 00:15:45.308 "data_size": 65536 00:15:45.308 } 00:15:45.308 ] 00:15:45.308 }' 00:15:45.308 15:21:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:45.308 15:21:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:45.568 15:21:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:15:45.568 15:21:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:45.569 15:21:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:45.569 15:21:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:45.569 15:21:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:45.569 15:21:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:15:45.569 15:21:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:45.569 15:21:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:15:45.569 15:21:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:45.569 15:21:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:45.569 15:21:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:45.569 15:21:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u bf21cf7a-ddbe-4728-9752-6bf57f2e0c1f 00:15:45.569 15:21:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:45.569 15:21:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:45.569 [2024-11-19 15:21:35.871511] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:15:45.569 [2024-11-19 15:21:35.871554] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:15:45.569 [2024-11-19 15:21:35.871561] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:15:45.569 [2024-11-19 15:21:35.871826] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:15:45.569 [2024-11-19 15:21:35.872293] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:15:45.569 [2024-11-19 15:21:35.872313] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:15:45.569 [2024-11-19 15:21:35.872473] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:45.569 NewBaseBdev 00:15:45.569 15:21:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:45.569 15:21:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:15:45.569 15:21:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:15:45.569 15:21:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:45.569 15:21:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:15:45.569 15:21:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:45.569 15:21:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:45.569 15:21:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:45.569 15:21:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:45.569 15:21:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:45.569 15:21:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:45.569 15:21:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:15:45.569 15:21:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:45.569 15:21:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:45.569 [ 00:15:45.569 { 00:15:45.569 "name": "NewBaseBdev", 00:15:45.569 "aliases": [ 00:15:45.569 "bf21cf7a-ddbe-4728-9752-6bf57f2e0c1f" 00:15:45.569 ], 00:15:45.569 "product_name": "Malloc disk", 00:15:45.569 "block_size": 512, 00:15:45.569 "num_blocks": 65536, 00:15:45.569 "uuid": "bf21cf7a-ddbe-4728-9752-6bf57f2e0c1f", 00:15:45.569 "assigned_rate_limits": { 00:15:45.569 "rw_ios_per_sec": 0, 00:15:45.569 "rw_mbytes_per_sec": 0, 00:15:45.569 "r_mbytes_per_sec": 0, 00:15:45.569 "w_mbytes_per_sec": 0 00:15:45.569 }, 00:15:45.569 "claimed": true, 00:15:45.569 "claim_type": "exclusive_write", 00:15:45.569 "zoned": false, 00:15:45.569 "supported_io_types": { 00:15:45.569 "read": true, 00:15:45.569 "write": true, 00:15:45.569 "unmap": true, 00:15:45.569 "flush": true, 00:15:45.569 "reset": true, 00:15:45.569 "nvme_admin": false, 00:15:45.569 "nvme_io": false, 00:15:45.569 "nvme_io_md": false, 00:15:45.569 "write_zeroes": true, 00:15:45.569 "zcopy": true, 00:15:45.569 "get_zone_info": false, 00:15:45.569 "zone_management": false, 00:15:45.569 "zone_append": false, 00:15:45.569 "compare": false, 00:15:45.829 "compare_and_write": false, 00:15:45.829 "abort": true, 00:15:45.829 "seek_hole": false, 00:15:45.829 "seek_data": false, 00:15:45.829 "copy": true, 00:15:45.829 "nvme_iov_md": false 00:15:45.829 }, 00:15:45.829 "memory_domains": [ 00:15:45.829 { 00:15:45.829 "dma_device_id": "system", 00:15:45.829 "dma_device_type": 1 00:15:45.829 }, 00:15:45.829 { 00:15:45.829 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:45.829 "dma_device_type": 2 00:15:45.829 } 00:15:45.829 ], 00:15:45.829 "driver_specific": {} 00:15:45.829 } 00:15:45.829 ] 00:15:45.829 15:21:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:45.829 15:21:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:15:45.829 15:21:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:15:45.829 15:21:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:45.829 15:21:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:45.829 15:21:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:45.829 15:21:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:45.829 15:21:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:45.829 15:21:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:45.829 15:21:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:45.829 15:21:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:45.829 15:21:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:45.829 15:21:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:45.829 15:21:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:45.829 15:21:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:45.829 15:21:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:45.829 15:21:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:45.829 15:21:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:45.829 "name": "Existed_Raid", 00:15:45.829 "uuid": "42cc0e88-5d21-488a-80b6-2c9dc1916876", 00:15:45.829 "strip_size_kb": 64, 00:15:45.829 "state": "online", 00:15:45.829 "raid_level": "raid5f", 00:15:45.829 "superblock": false, 00:15:45.829 "num_base_bdevs": 4, 00:15:45.829 "num_base_bdevs_discovered": 4, 00:15:45.829 "num_base_bdevs_operational": 4, 00:15:45.829 "base_bdevs_list": [ 00:15:45.829 { 00:15:45.829 "name": "NewBaseBdev", 00:15:45.829 "uuid": "bf21cf7a-ddbe-4728-9752-6bf57f2e0c1f", 00:15:45.829 "is_configured": true, 00:15:45.829 "data_offset": 0, 00:15:45.829 "data_size": 65536 00:15:45.829 }, 00:15:45.829 { 00:15:45.829 "name": "BaseBdev2", 00:15:45.829 "uuid": "03a46437-56c7-47ec-82ff-d54f5d087971", 00:15:45.829 "is_configured": true, 00:15:45.829 "data_offset": 0, 00:15:45.829 "data_size": 65536 00:15:45.829 }, 00:15:45.829 { 00:15:45.829 "name": "BaseBdev3", 00:15:45.829 "uuid": "f846b0df-70a7-44e6-9dfa-d0d03d606d82", 00:15:45.829 "is_configured": true, 00:15:45.829 "data_offset": 0, 00:15:45.829 "data_size": 65536 00:15:45.829 }, 00:15:45.829 { 00:15:45.829 "name": "BaseBdev4", 00:15:45.829 "uuid": "a72dbccf-b414-4e92-9f37-af8f4174426f", 00:15:45.829 "is_configured": true, 00:15:45.829 "data_offset": 0, 00:15:45.829 "data_size": 65536 00:15:45.829 } 00:15:45.829 ] 00:15:45.829 }' 00:15:45.829 15:21:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:45.829 15:21:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:46.089 15:21:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:15:46.089 15:21:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:15:46.089 15:21:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:46.089 15:21:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:46.089 15:21:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:15:46.089 15:21:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:46.089 15:21:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:46.089 15:21:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:15:46.089 15:21:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:46.089 15:21:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:46.089 [2024-11-19 15:21:36.334922] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:46.089 15:21:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:46.089 15:21:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:46.089 "name": "Existed_Raid", 00:15:46.089 "aliases": [ 00:15:46.089 "42cc0e88-5d21-488a-80b6-2c9dc1916876" 00:15:46.089 ], 00:15:46.089 "product_name": "Raid Volume", 00:15:46.089 "block_size": 512, 00:15:46.089 "num_blocks": 196608, 00:15:46.089 "uuid": "42cc0e88-5d21-488a-80b6-2c9dc1916876", 00:15:46.089 "assigned_rate_limits": { 00:15:46.089 "rw_ios_per_sec": 0, 00:15:46.089 "rw_mbytes_per_sec": 0, 00:15:46.089 "r_mbytes_per_sec": 0, 00:15:46.089 "w_mbytes_per_sec": 0 00:15:46.089 }, 00:15:46.089 "claimed": false, 00:15:46.089 "zoned": false, 00:15:46.089 "supported_io_types": { 00:15:46.089 "read": true, 00:15:46.089 "write": true, 00:15:46.089 "unmap": false, 00:15:46.089 "flush": false, 00:15:46.090 "reset": true, 00:15:46.090 "nvme_admin": false, 00:15:46.090 "nvme_io": false, 00:15:46.090 "nvme_io_md": false, 00:15:46.090 "write_zeroes": true, 00:15:46.090 "zcopy": false, 00:15:46.090 "get_zone_info": false, 00:15:46.090 "zone_management": false, 00:15:46.090 "zone_append": false, 00:15:46.090 "compare": false, 00:15:46.090 "compare_and_write": false, 00:15:46.090 "abort": false, 00:15:46.090 "seek_hole": false, 00:15:46.090 "seek_data": false, 00:15:46.090 "copy": false, 00:15:46.090 "nvme_iov_md": false 00:15:46.090 }, 00:15:46.090 "driver_specific": { 00:15:46.090 "raid": { 00:15:46.090 "uuid": "42cc0e88-5d21-488a-80b6-2c9dc1916876", 00:15:46.090 "strip_size_kb": 64, 00:15:46.090 "state": "online", 00:15:46.090 "raid_level": "raid5f", 00:15:46.090 "superblock": false, 00:15:46.090 "num_base_bdevs": 4, 00:15:46.090 "num_base_bdevs_discovered": 4, 00:15:46.090 "num_base_bdevs_operational": 4, 00:15:46.090 "base_bdevs_list": [ 00:15:46.090 { 00:15:46.090 "name": "NewBaseBdev", 00:15:46.090 "uuid": "bf21cf7a-ddbe-4728-9752-6bf57f2e0c1f", 00:15:46.090 "is_configured": true, 00:15:46.090 "data_offset": 0, 00:15:46.090 "data_size": 65536 00:15:46.090 }, 00:15:46.090 { 00:15:46.090 "name": "BaseBdev2", 00:15:46.090 "uuid": "03a46437-56c7-47ec-82ff-d54f5d087971", 00:15:46.090 "is_configured": true, 00:15:46.090 "data_offset": 0, 00:15:46.090 "data_size": 65536 00:15:46.090 }, 00:15:46.090 { 00:15:46.090 "name": "BaseBdev3", 00:15:46.090 "uuid": "f846b0df-70a7-44e6-9dfa-d0d03d606d82", 00:15:46.090 "is_configured": true, 00:15:46.090 "data_offset": 0, 00:15:46.090 "data_size": 65536 00:15:46.090 }, 00:15:46.090 { 00:15:46.090 "name": "BaseBdev4", 00:15:46.090 "uuid": "a72dbccf-b414-4e92-9f37-af8f4174426f", 00:15:46.090 "is_configured": true, 00:15:46.090 "data_offset": 0, 00:15:46.090 "data_size": 65536 00:15:46.090 } 00:15:46.090 ] 00:15:46.090 } 00:15:46.090 } 00:15:46.090 }' 00:15:46.090 15:21:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:46.090 15:21:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:15:46.090 BaseBdev2 00:15:46.090 BaseBdev3 00:15:46.090 BaseBdev4' 00:15:46.090 15:21:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:46.350 15:21:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:15:46.350 15:21:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:46.350 15:21:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:15:46.350 15:21:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:46.350 15:21:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:46.350 15:21:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:46.350 15:21:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:46.350 15:21:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:46.350 15:21:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:46.350 15:21:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:46.350 15:21:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:15:46.350 15:21:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:46.350 15:21:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:46.350 15:21:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:46.350 15:21:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:46.350 15:21:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:46.350 15:21:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:46.350 15:21:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:46.350 15:21:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:46.350 15:21:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:15:46.350 15:21:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:46.350 15:21:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:46.350 15:21:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:46.350 15:21:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:46.350 15:21:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:46.350 15:21:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:46.350 15:21:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:15:46.350 15:21:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:46.350 15:21:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:46.350 15:21:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:46.350 15:21:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:46.350 15:21:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:46.350 15:21:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:46.350 15:21:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:15:46.350 15:21:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:46.350 15:21:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:46.350 [2024-11-19 15:21:36.662190] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:15:46.350 [2024-11-19 15:21:36.662216] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:46.350 [2024-11-19 15:21:36.662281] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:46.350 [2024-11-19 15:21:36.662521] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:46.350 [2024-11-19 15:21:36.662532] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:15:46.350 15:21:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:46.350 15:21:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 93252 00:15:46.350 15:21:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 93252 ']' 00:15:46.350 15:21:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@958 -- # kill -0 93252 00:15:46.350 15:21:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@959 -- # uname 00:15:46.350 15:21:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:15:46.350 15:21:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 93252 00:15:46.610 15:21:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:15:46.610 15:21:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:15:46.610 15:21:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 93252' 00:15:46.610 killing process with pid 93252 00:15:46.610 15:21:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@973 -- # kill 93252 00:15:46.610 [2024-11-19 15:21:36.712328] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:46.610 15:21:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@978 -- # wait 93252 00:15:46.610 [2024-11-19 15:21:36.754291] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:46.871 15:21:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:15:46.871 00:15:46.871 real 0m9.720s 00:15:46.871 user 0m16.586s 00:15:46.871 sys 0m2.191s 00:15:46.871 15:21:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:15:46.871 15:21:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:15:46.871 ************************************ 00:15:46.871 END TEST raid5f_state_function_test 00:15:46.871 ************************************ 00:15:46.871 15:21:37 bdev_raid -- bdev/bdev_raid.sh@987 -- # run_test raid5f_state_function_test_sb raid_state_function_test raid5f 4 true 00:15:46.871 15:21:37 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:15:46.871 15:21:37 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:15:46.871 15:21:37 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:46.871 ************************************ 00:15:46.871 START TEST raid5f_state_function_test_sb 00:15:46.871 ************************************ 00:15:46.871 15:21:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid5f 4 true 00:15:46.871 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:15:46.871 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:15:46.871 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:15:46.871 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:15:46.871 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:15:46.871 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:46.871 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:15:46.871 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:15:46.871 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:46.871 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:15:46.871 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:15:46.871 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:46.871 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:15:46.871 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:15:46.871 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:46.871 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:15:46.871 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:15:46.871 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:46.871 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:15:46.871 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:15:46.871 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:15:46.871 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:15:46.871 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:15:46.871 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:15:46.871 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:15:46.871 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:15:46.871 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:15:46.871 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:15:46.871 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:15:46.871 Process raid pid: 93907 00:15:46.871 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=93907 00:15:46.871 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:15:46.871 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 93907' 00:15:46.871 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 93907 00:15:46.871 15:21:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 93907 ']' 00:15:46.871 15:21:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:46.871 15:21:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:15:46.871 15:21:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:46.871 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:46.871 15:21:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:15:46.871 15:21:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:46.871 [2024-11-19 15:21:37.153408] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:15:46.871 [2024-11-19 15:21:37.153618] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:15:47.130 [2024-11-19 15:21:37.310032] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:47.131 [2024-11-19 15:21:37.335325] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:47.131 [2024-11-19 15:21:37.380166] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:47.131 [2024-11-19 15:21:37.380272] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:47.700 15:21:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:15:47.700 15:21:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:15:47.700 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:15:47.700 15:21:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:47.700 15:21:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:47.700 [2024-11-19 15:21:37.950235] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:15:47.700 [2024-11-19 15:21:37.950347] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:15:47.700 [2024-11-19 15:21:37.950387] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:47.700 [2024-11-19 15:21:37.950410] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:47.700 [2024-11-19 15:21:37.950428] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:15:47.700 [2024-11-19 15:21:37.950450] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:15:47.700 [2024-11-19 15:21:37.950472] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:15:47.700 [2024-11-19 15:21:37.950495] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:15:47.700 15:21:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:47.700 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:47.700 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:47.700 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:47.700 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:47.700 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:47.700 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:47.700 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:47.700 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:47.700 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:47.700 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:47.700 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:47.700 15:21:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:47.700 15:21:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:47.700 15:21:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:47.700 15:21:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:47.700 15:21:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:47.700 "name": "Existed_Raid", 00:15:47.700 "uuid": "19cc49f0-7380-4f9d-bb58-6cb887a32566", 00:15:47.700 "strip_size_kb": 64, 00:15:47.700 "state": "configuring", 00:15:47.700 "raid_level": "raid5f", 00:15:47.700 "superblock": true, 00:15:47.700 "num_base_bdevs": 4, 00:15:47.700 "num_base_bdevs_discovered": 0, 00:15:47.700 "num_base_bdevs_operational": 4, 00:15:47.700 "base_bdevs_list": [ 00:15:47.700 { 00:15:47.700 "name": "BaseBdev1", 00:15:47.700 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:47.700 "is_configured": false, 00:15:47.700 "data_offset": 0, 00:15:47.700 "data_size": 0 00:15:47.700 }, 00:15:47.700 { 00:15:47.700 "name": "BaseBdev2", 00:15:47.700 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:47.700 "is_configured": false, 00:15:47.700 "data_offset": 0, 00:15:47.700 "data_size": 0 00:15:47.700 }, 00:15:47.700 { 00:15:47.700 "name": "BaseBdev3", 00:15:47.700 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:47.700 "is_configured": false, 00:15:47.700 "data_offset": 0, 00:15:47.700 "data_size": 0 00:15:47.700 }, 00:15:47.700 { 00:15:47.700 "name": "BaseBdev4", 00:15:47.700 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:47.700 "is_configured": false, 00:15:47.700 "data_offset": 0, 00:15:47.700 "data_size": 0 00:15:47.700 } 00:15:47.700 ] 00:15:47.700 }' 00:15:47.700 15:21:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:47.700 15:21:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:48.270 [2024-11-19 15:21:38.437281] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:15:48.270 [2024-11-19 15:21:38.437351] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:48.270 [2024-11-19 15:21:38.449294] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:15:48.270 [2024-11-19 15:21:38.449379] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:15:48.270 [2024-11-19 15:21:38.449404] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:48.270 [2024-11-19 15:21:38.449425] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:48.270 [2024-11-19 15:21:38.449442] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:15:48.270 [2024-11-19 15:21:38.449462] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:15:48.270 [2024-11-19 15:21:38.449478] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:15:48.270 [2024-11-19 15:21:38.449497] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:48.270 [2024-11-19 15:21:38.470350] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:48.270 BaseBdev1 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:48.270 [ 00:15:48.270 { 00:15:48.270 "name": "BaseBdev1", 00:15:48.270 "aliases": [ 00:15:48.270 "8f7afaf9-b4ee-4b6b-bc84-7127d8f99045" 00:15:48.270 ], 00:15:48.270 "product_name": "Malloc disk", 00:15:48.270 "block_size": 512, 00:15:48.270 "num_blocks": 65536, 00:15:48.270 "uuid": "8f7afaf9-b4ee-4b6b-bc84-7127d8f99045", 00:15:48.270 "assigned_rate_limits": { 00:15:48.270 "rw_ios_per_sec": 0, 00:15:48.270 "rw_mbytes_per_sec": 0, 00:15:48.270 "r_mbytes_per_sec": 0, 00:15:48.270 "w_mbytes_per_sec": 0 00:15:48.270 }, 00:15:48.270 "claimed": true, 00:15:48.270 "claim_type": "exclusive_write", 00:15:48.270 "zoned": false, 00:15:48.270 "supported_io_types": { 00:15:48.270 "read": true, 00:15:48.270 "write": true, 00:15:48.270 "unmap": true, 00:15:48.270 "flush": true, 00:15:48.270 "reset": true, 00:15:48.270 "nvme_admin": false, 00:15:48.270 "nvme_io": false, 00:15:48.270 "nvme_io_md": false, 00:15:48.270 "write_zeroes": true, 00:15:48.270 "zcopy": true, 00:15:48.270 "get_zone_info": false, 00:15:48.270 "zone_management": false, 00:15:48.270 "zone_append": false, 00:15:48.270 "compare": false, 00:15:48.270 "compare_and_write": false, 00:15:48.270 "abort": true, 00:15:48.270 "seek_hole": false, 00:15:48.270 "seek_data": false, 00:15:48.270 "copy": true, 00:15:48.270 "nvme_iov_md": false 00:15:48.270 }, 00:15:48.270 "memory_domains": [ 00:15:48.270 { 00:15:48.270 "dma_device_id": "system", 00:15:48.270 "dma_device_type": 1 00:15:48.270 }, 00:15:48.270 { 00:15:48.270 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:48.270 "dma_device_type": 2 00:15:48.270 } 00:15:48.270 ], 00:15:48.270 "driver_specific": {} 00:15:48.270 } 00:15:48.270 ] 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:48.270 "name": "Existed_Raid", 00:15:48.270 "uuid": "273a71fd-f72d-40b2-add4-dff5db8d6b6d", 00:15:48.270 "strip_size_kb": 64, 00:15:48.270 "state": "configuring", 00:15:48.270 "raid_level": "raid5f", 00:15:48.270 "superblock": true, 00:15:48.270 "num_base_bdevs": 4, 00:15:48.270 "num_base_bdevs_discovered": 1, 00:15:48.270 "num_base_bdevs_operational": 4, 00:15:48.270 "base_bdevs_list": [ 00:15:48.270 { 00:15:48.270 "name": "BaseBdev1", 00:15:48.270 "uuid": "8f7afaf9-b4ee-4b6b-bc84-7127d8f99045", 00:15:48.270 "is_configured": true, 00:15:48.270 "data_offset": 2048, 00:15:48.270 "data_size": 63488 00:15:48.270 }, 00:15:48.270 { 00:15:48.270 "name": "BaseBdev2", 00:15:48.270 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:48.270 "is_configured": false, 00:15:48.270 "data_offset": 0, 00:15:48.270 "data_size": 0 00:15:48.270 }, 00:15:48.270 { 00:15:48.270 "name": "BaseBdev3", 00:15:48.270 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:48.270 "is_configured": false, 00:15:48.270 "data_offset": 0, 00:15:48.270 "data_size": 0 00:15:48.270 }, 00:15:48.270 { 00:15:48.270 "name": "BaseBdev4", 00:15:48.270 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:48.270 "is_configured": false, 00:15:48.270 "data_offset": 0, 00:15:48.270 "data_size": 0 00:15:48.270 } 00:15:48.270 ] 00:15:48.270 }' 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:48.270 15:21:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:48.841 15:21:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:15:48.841 15:21:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:48.841 15:21:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:48.841 [2024-11-19 15:21:38.973501] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:15:48.841 [2024-11-19 15:21:38.973543] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:15:48.841 15:21:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:48.841 15:21:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:15:48.841 15:21:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:48.841 15:21:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:48.841 [2024-11-19 15:21:38.985525] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:48.841 [2024-11-19 15:21:38.987316] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:48.841 [2024-11-19 15:21:38.987355] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:48.841 [2024-11-19 15:21:38.987364] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:15:48.841 [2024-11-19 15:21:38.987373] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:15:48.841 [2024-11-19 15:21:38.987379] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:15:48.841 [2024-11-19 15:21:38.987387] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:15:48.841 15:21:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:48.841 15:21:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:15:48.841 15:21:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:15:48.841 15:21:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:48.841 15:21:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:48.841 15:21:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:48.841 15:21:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:48.841 15:21:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:48.841 15:21:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:48.841 15:21:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:48.841 15:21:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:48.841 15:21:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:48.841 15:21:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:48.841 15:21:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:48.841 15:21:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:48.841 15:21:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:48.841 15:21:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:48.841 15:21:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:48.841 15:21:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:48.841 "name": "Existed_Raid", 00:15:48.841 "uuid": "df0b7e60-da0c-4f86-ad98-f67097e5997e", 00:15:48.841 "strip_size_kb": 64, 00:15:48.841 "state": "configuring", 00:15:48.841 "raid_level": "raid5f", 00:15:48.841 "superblock": true, 00:15:48.841 "num_base_bdevs": 4, 00:15:48.841 "num_base_bdevs_discovered": 1, 00:15:48.841 "num_base_bdevs_operational": 4, 00:15:48.841 "base_bdevs_list": [ 00:15:48.841 { 00:15:48.841 "name": "BaseBdev1", 00:15:48.841 "uuid": "8f7afaf9-b4ee-4b6b-bc84-7127d8f99045", 00:15:48.841 "is_configured": true, 00:15:48.841 "data_offset": 2048, 00:15:48.841 "data_size": 63488 00:15:48.841 }, 00:15:48.841 { 00:15:48.841 "name": "BaseBdev2", 00:15:48.841 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:48.841 "is_configured": false, 00:15:48.841 "data_offset": 0, 00:15:48.841 "data_size": 0 00:15:48.841 }, 00:15:48.841 { 00:15:48.841 "name": "BaseBdev3", 00:15:48.841 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:48.841 "is_configured": false, 00:15:48.841 "data_offset": 0, 00:15:48.841 "data_size": 0 00:15:48.841 }, 00:15:48.841 { 00:15:48.841 "name": "BaseBdev4", 00:15:48.841 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:48.841 "is_configured": false, 00:15:48.841 "data_offset": 0, 00:15:48.841 "data_size": 0 00:15:48.841 } 00:15:48.841 ] 00:15:48.841 }' 00:15:48.841 15:21:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:48.841 15:21:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:49.100 15:21:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:15:49.100 15:21:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:49.100 15:21:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:49.100 [2024-11-19 15:21:39.419716] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:49.100 BaseBdev2 00:15:49.100 15:21:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:49.100 15:21:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:15:49.100 15:21:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:15:49.100 15:21:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:49.100 15:21:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:15:49.100 15:21:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:49.100 15:21:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:49.100 15:21:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:49.100 15:21:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:49.100 15:21:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:49.100 15:21:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:49.100 15:21:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:15:49.100 15:21:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:49.100 15:21:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:49.360 [ 00:15:49.360 { 00:15:49.360 "name": "BaseBdev2", 00:15:49.360 "aliases": [ 00:15:49.360 "cd3252a8-7e38-4641-b880-26bbe291d8b7" 00:15:49.360 ], 00:15:49.360 "product_name": "Malloc disk", 00:15:49.360 "block_size": 512, 00:15:49.360 "num_blocks": 65536, 00:15:49.360 "uuid": "cd3252a8-7e38-4641-b880-26bbe291d8b7", 00:15:49.360 "assigned_rate_limits": { 00:15:49.360 "rw_ios_per_sec": 0, 00:15:49.360 "rw_mbytes_per_sec": 0, 00:15:49.360 "r_mbytes_per_sec": 0, 00:15:49.360 "w_mbytes_per_sec": 0 00:15:49.360 }, 00:15:49.360 "claimed": true, 00:15:49.360 "claim_type": "exclusive_write", 00:15:49.360 "zoned": false, 00:15:49.360 "supported_io_types": { 00:15:49.360 "read": true, 00:15:49.360 "write": true, 00:15:49.360 "unmap": true, 00:15:49.360 "flush": true, 00:15:49.360 "reset": true, 00:15:49.360 "nvme_admin": false, 00:15:49.360 "nvme_io": false, 00:15:49.360 "nvme_io_md": false, 00:15:49.360 "write_zeroes": true, 00:15:49.360 "zcopy": true, 00:15:49.360 "get_zone_info": false, 00:15:49.360 "zone_management": false, 00:15:49.360 "zone_append": false, 00:15:49.360 "compare": false, 00:15:49.360 "compare_and_write": false, 00:15:49.360 "abort": true, 00:15:49.360 "seek_hole": false, 00:15:49.360 "seek_data": false, 00:15:49.360 "copy": true, 00:15:49.360 "nvme_iov_md": false 00:15:49.360 }, 00:15:49.360 "memory_domains": [ 00:15:49.360 { 00:15:49.360 "dma_device_id": "system", 00:15:49.361 "dma_device_type": 1 00:15:49.361 }, 00:15:49.361 { 00:15:49.361 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:49.361 "dma_device_type": 2 00:15:49.361 } 00:15:49.361 ], 00:15:49.361 "driver_specific": {} 00:15:49.361 } 00:15:49.361 ] 00:15:49.361 15:21:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:49.361 15:21:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:15:49.361 15:21:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:15:49.361 15:21:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:15:49.361 15:21:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:49.361 15:21:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:49.361 15:21:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:49.361 15:21:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:49.361 15:21:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:49.361 15:21:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:49.361 15:21:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:49.361 15:21:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:49.361 15:21:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:49.361 15:21:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:49.361 15:21:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:49.361 15:21:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:49.361 15:21:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:49.361 15:21:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:49.361 15:21:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:49.361 15:21:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:49.361 "name": "Existed_Raid", 00:15:49.361 "uuid": "df0b7e60-da0c-4f86-ad98-f67097e5997e", 00:15:49.361 "strip_size_kb": 64, 00:15:49.361 "state": "configuring", 00:15:49.361 "raid_level": "raid5f", 00:15:49.361 "superblock": true, 00:15:49.361 "num_base_bdevs": 4, 00:15:49.361 "num_base_bdevs_discovered": 2, 00:15:49.361 "num_base_bdevs_operational": 4, 00:15:49.361 "base_bdevs_list": [ 00:15:49.361 { 00:15:49.361 "name": "BaseBdev1", 00:15:49.361 "uuid": "8f7afaf9-b4ee-4b6b-bc84-7127d8f99045", 00:15:49.361 "is_configured": true, 00:15:49.361 "data_offset": 2048, 00:15:49.361 "data_size": 63488 00:15:49.361 }, 00:15:49.361 { 00:15:49.361 "name": "BaseBdev2", 00:15:49.361 "uuid": "cd3252a8-7e38-4641-b880-26bbe291d8b7", 00:15:49.361 "is_configured": true, 00:15:49.361 "data_offset": 2048, 00:15:49.361 "data_size": 63488 00:15:49.361 }, 00:15:49.361 { 00:15:49.361 "name": "BaseBdev3", 00:15:49.361 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:49.361 "is_configured": false, 00:15:49.361 "data_offset": 0, 00:15:49.361 "data_size": 0 00:15:49.361 }, 00:15:49.361 { 00:15:49.361 "name": "BaseBdev4", 00:15:49.361 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:49.361 "is_configured": false, 00:15:49.361 "data_offset": 0, 00:15:49.361 "data_size": 0 00:15:49.361 } 00:15:49.361 ] 00:15:49.361 }' 00:15:49.361 15:21:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:49.361 15:21:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:49.622 15:21:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:15:49.622 15:21:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:49.622 15:21:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:49.622 [2024-11-19 15:21:39.910915] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:15:49.622 BaseBdev3 00:15:49.622 15:21:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:49.622 15:21:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:15:49.622 15:21:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:15:49.622 15:21:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:49.622 15:21:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:15:49.622 15:21:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:49.622 15:21:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:49.622 15:21:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:49.622 15:21:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:49.622 15:21:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:49.622 15:21:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:49.622 15:21:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:15:49.622 15:21:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:49.622 15:21:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:49.622 [ 00:15:49.622 { 00:15:49.622 "name": "BaseBdev3", 00:15:49.622 "aliases": [ 00:15:49.622 "57065316-d42e-4947-93b5-56f213c9342f" 00:15:49.622 ], 00:15:49.622 "product_name": "Malloc disk", 00:15:49.622 "block_size": 512, 00:15:49.622 "num_blocks": 65536, 00:15:49.622 "uuid": "57065316-d42e-4947-93b5-56f213c9342f", 00:15:49.622 "assigned_rate_limits": { 00:15:49.622 "rw_ios_per_sec": 0, 00:15:49.622 "rw_mbytes_per_sec": 0, 00:15:49.622 "r_mbytes_per_sec": 0, 00:15:49.622 "w_mbytes_per_sec": 0 00:15:49.622 }, 00:15:49.622 "claimed": true, 00:15:49.622 "claim_type": "exclusive_write", 00:15:49.622 "zoned": false, 00:15:49.622 "supported_io_types": { 00:15:49.622 "read": true, 00:15:49.622 "write": true, 00:15:49.622 "unmap": true, 00:15:49.622 "flush": true, 00:15:49.622 "reset": true, 00:15:49.622 "nvme_admin": false, 00:15:49.622 "nvme_io": false, 00:15:49.622 "nvme_io_md": false, 00:15:49.622 "write_zeroes": true, 00:15:49.622 "zcopy": true, 00:15:49.622 "get_zone_info": false, 00:15:49.622 "zone_management": false, 00:15:49.622 "zone_append": false, 00:15:49.622 "compare": false, 00:15:49.622 "compare_and_write": false, 00:15:49.622 "abort": true, 00:15:49.622 "seek_hole": false, 00:15:49.622 "seek_data": false, 00:15:49.622 "copy": true, 00:15:49.622 "nvme_iov_md": false 00:15:49.622 }, 00:15:49.622 "memory_domains": [ 00:15:49.622 { 00:15:49.622 "dma_device_id": "system", 00:15:49.622 "dma_device_type": 1 00:15:49.622 }, 00:15:49.622 { 00:15:49.622 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:49.622 "dma_device_type": 2 00:15:49.622 } 00:15:49.622 ], 00:15:49.622 "driver_specific": {} 00:15:49.622 } 00:15:49.622 ] 00:15:49.622 15:21:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:49.622 15:21:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:15:49.622 15:21:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:15:49.622 15:21:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:15:49.622 15:21:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:49.622 15:21:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:49.622 15:21:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:49.622 15:21:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:49.622 15:21:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:49.622 15:21:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:49.622 15:21:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:49.622 15:21:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:49.622 15:21:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:49.622 15:21:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:49.622 15:21:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:49.622 15:21:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:49.622 15:21:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:49.622 15:21:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:49.882 15:21:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:49.882 15:21:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:49.882 "name": "Existed_Raid", 00:15:49.882 "uuid": "df0b7e60-da0c-4f86-ad98-f67097e5997e", 00:15:49.882 "strip_size_kb": 64, 00:15:49.882 "state": "configuring", 00:15:49.882 "raid_level": "raid5f", 00:15:49.882 "superblock": true, 00:15:49.882 "num_base_bdevs": 4, 00:15:49.882 "num_base_bdevs_discovered": 3, 00:15:49.882 "num_base_bdevs_operational": 4, 00:15:49.882 "base_bdevs_list": [ 00:15:49.882 { 00:15:49.882 "name": "BaseBdev1", 00:15:49.882 "uuid": "8f7afaf9-b4ee-4b6b-bc84-7127d8f99045", 00:15:49.882 "is_configured": true, 00:15:49.882 "data_offset": 2048, 00:15:49.882 "data_size": 63488 00:15:49.882 }, 00:15:49.882 { 00:15:49.882 "name": "BaseBdev2", 00:15:49.882 "uuid": "cd3252a8-7e38-4641-b880-26bbe291d8b7", 00:15:49.882 "is_configured": true, 00:15:49.882 "data_offset": 2048, 00:15:49.882 "data_size": 63488 00:15:49.882 }, 00:15:49.882 { 00:15:49.882 "name": "BaseBdev3", 00:15:49.882 "uuid": "57065316-d42e-4947-93b5-56f213c9342f", 00:15:49.882 "is_configured": true, 00:15:49.882 "data_offset": 2048, 00:15:49.882 "data_size": 63488 00:15:49.882 }, 00:15:49.882 { 00:15:49.882 "name": "BaseBdev4", 00:15:49.882 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:49.882 "is_configured": false, 00:15:49.882 "data_offset": 0, 00:15:49.882 "data_size": 0 00:15:49.882 } 00:15:49.882 ] 00:15:49.882 }' 00:15:49.882 15:21:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:49.882 15:21:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:50.143 15:21:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:15:50.143 15:21:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:50.143 15:21:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:50.143 [2024-11-19 15:21:40.432909] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:15:50.143 [2024-11-19 15:21:40.433201] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:15:50.143 [2024-11-19 15:21:40.433252] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:15:50.143 [2024-11-19 15:21:40.433536] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:15:50.143 BaseBdev4 00:15:50.143 [2024-11-19 15:21:40.434075] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:15:50.143 [2024-11-19 15:21:40.434132] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:15:50.143 [2024-11-19 15:21:40.434288] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:50.143 15:21:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:50.143 15:21:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:15:50.143 15:21:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:15:50.143 15:21:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:50.143 15:21:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:15:50.143 15:21:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:50.143 15:21:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:50.143 15:21:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:50.143 15:21:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:50.143 15:21:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:50.143 15:21:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:50.143 15:21:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:15:50.143 15:21:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:50.143 15:21:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:50.143 [ 00:15:50.143 { 00:15:50.143 "name": "BaseBdev4", 00:15:50.143 "aliases": [ 00:15:50.143 "28c942c6-cb9f-4847-8070-e2f428e7c321" 00:15:50.143 ], 00:15:50.143 "product_name": "Malloc disk", 00:15:50.143 "block_size": 512, 00:15:50.143 "num_blocks": 65536, 00:15:50.143 "uuid": "28c942c6-cb9f-4847-8070-e2f428e7c321", 00:15:50.143 "assigned_rate_limits": { 00:15:50.143 "rw_ios_per_sec": 0, 00:15:50.143 "rw_mbytes_per_sec": 0, 00:15:50.143 "r_mbytes_per_sec": 0, 00:15:50.143 "w_mbytes_per_sec": 0 00:15:50.143 }, 00:15:50.143 "claimed": true, 00:15:50.143 "claim_type": "exclusive_write", 00:15:50.143 "zoned": false, 00:15:50.143 "supported_io_types": { 00:15:50.143 "read": true, 00:15:50.143 "write": true, 00:15:50.143 "unmap": true, 00:15:50.143 "flush": true, 00:15:50.143 "reset": true, 00:15:50.143 "nvme_admin": false, 00:15:50.143 "nvme_io": false, 00:15:50.143 "nvme_io_md": false, 00:15:50.143 "write_zeroes": true, 00:15:50.143 "zcopy": true, 00:15:50.143 "get_zone_info": false, 00:15:50.143 "zone_management": false, 00:15:50.143 "zone_append": false, 00:15:50.143 "compare": false, 00:15:50.143 "compare_and_write": false, 00:15:50.143 "abort": true, 00:15:50.143 "seek_hole": false, 00:15:50.143 "seek_data": false, 00:15:50.143 "copy": true, 00:15:50.143 "nvme_iov_md": false 00:15:50.143 }, 00:15:50.143 "memory_domains": [ 00:15:50.143 { 00:15:50.143 "dma_device_id": "system", 00:15:50.143 "dma_device_type": 1 00:15:50.143 }, 00:15:50.143 { 00:15:50.143 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:50.143 "dma_device_type": 2 00:15:50.143 } 00:15:50.143 ], 00:15:50.143 "driver_specific": {} 00:15:50.143 } 00:15:50.143 ] 00:15:50.143 15:21:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:50.143 15:21:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:15:50.143 15:21:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:15:50.143 15:21:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:15:50.143 15:21:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:15:50.143 15:21:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:50.143 15:21:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:50.143 15:21:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:50.143 15:21:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:50.143 15:21:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:50.143 15:21:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:50.143 15:21:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:50.143 15:21:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:50.143 15:21:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:50.143 15:21:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:50.403 15:21:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:50.403 15:21:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:50.403 15:21:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:50.403 15:21:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:50.403 15:21:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:50.403 "name": "Existed_Raid", 00:15:50.403 "uuid": "df0b7e60-da0c-4f86-ad98-f67097e5997e", 00:15:50.403 "strip_size_kb": 64, 00:15:50.403 "state": "online", 00:15:50.403 "raid_level": "raid5f", 00:15:50.403 "superblock": true, 00:15:50.403 "num_base_bdevs": 4, 00:15:50.403 "num_base_bdevs_discovered": 4, 00:15:50.403 "num_base_bdevs_operational": 4, 00:15:50.403 "base_bdevs_list": [ 00:15:50.403 { 00:15:50.403 "name": "BaseBdev1", 00:15:50.403 "uuid": "8f7afaf9-b4ee-4b6b-bc84-7127d8f99045", 00:15:50.403 "is_configured": true, 00:15:50.403 "data_offset": 2048, 00:15:50.403 "data_size": 63488 00:15:50.403 }, 00:15:50.403 { 00:15:50.403 "name": "BaseBdev2", 00:15:50.403 "uuid": "cd3252a8-7e38-4641-b880-26bbe291d8b7", 00:15:50.403 "is_configured": true, 00:15:50.403 "data_offset": 2048, 00:15:50.403 "data_size": 63488 00:15:50.403 }, 00:15:50.403 { 00:15:50.403 "name": "BaseBdev3", 00:15:50.403 "uuid": "57065316-d42e-4947-93b5-56f213c9342f", 00:15:50.403 "is_configured": true, 00:15:50.403 "data_offset": 2048, 00:15:50.403 "data_size": 63488 00:15:50.403 }, 00:15:50.403 { 00:15:50.403 "name": "BaseBdev4", 00:15:50.403 "uuid": "28c942c6-cb9f-4847-8070-e2f428e7c321", 00:15:50.403 "is_configured": true, 00:15:50.403 "data_offset": 2048, 00:15:50.403 "data_size": 63488 00:15:50.403 } 00:15:50.404 ] 00:15:50.404 }' 00:15:50.404 15:21:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:50.404 15:21:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:50.664 15:21:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:15:50.664 15:21:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:15:50.664 15:21:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:50.664 15:21:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:50.664 15:21:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:15:50.664 15:21:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:50.664 15:21:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:15:50.664 15:21:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:50.664 15:21:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:50.664 15:21:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:50.664 [2024-11-19 15:21:40.860426] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:50.664 15:21:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:50.664 15:21:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:50.664 "name": "Existed_Raid", 00:15:50.664 "aliases": [ 00:15:50.664 "df0b7e60-da0c-4f86-ad98-f67097e5997e" 00:15:50.664 ], 00:15:50.664 "product_name": "Raid Volume", 00:15:50.664 "block_size": 512, 00:15:50.664 "num_blocks": 190464, 00:15:50.664 "uuid": "df0b7e60-da0c-4f86-ad98-f67097e5997e", 00:15:50.664 "assigned_rate_limits": { 00:15:50.664 "rw_ios_per_sec": 0, 00:15:50.664 "rw_mbytes_per_sec": 0, 00:15:50.664 "r_mbytes_per_sec": 0, 00:15:50.664 "w_mbytes_per_sec": 0 00:15:50.664 }, 00:15:50.664 "claimed": false, 00:15:50.664 "zoned": false, 00:15:50.664 "supported_io_types": { 00:15:50.664 "read": true, 00:15:50.664 "write": true, 00:15:50.664 "unmap": false, 00:15:50.664 "flush": false, 00:15:50.664 "reset": true, 00:15:50.664 "nvme_admin": false, 00:15:50.664 "nvme_io": false, 00:15:50.664 "nvme_io_md": false, 00:15:50.664 "write_zeroes": true, 00:15:50.664 "zcopy": false, 00:15:50.664 "get_zone_info": false, 00:15:50.664 "zone_management": false, 00:15:50.664 "zone_append": false, 00:15:50.664 "compare": false, 00:15:50.664 "compare_and_write": false, 00:15:50.664 "abort": false, 00:15:50.664 "seek_hole": false, 00:15:50.664 "seek_data": false, 00:15:50.664 "copy": false, 00:15:50.664 "nvme_iov_md": false 00:15:50.664 }, 00:15:50.664 "driver_specific": { 00:15:50.664 "raid": { 00:15:50.664 "uuid": "df0b7e60-da0c-4f86-ad98-f67097e5997e", 00:15:50.664 "strip_size_kb": 64, 00:15:50.664 "state": "online", 00:15:50.664 "raid_level": "raid5f", 00:15:50.664 "superblock": true, 00:15:50.664 "num_base_bdevs": 4, 00:15:50.664 "num_base_bdevs_discovered": 4, 00:15:50.664 "num_base_bdevs_operational": 4, 00:15:50.664 "base_bdevs_list": [ 00:15:50.664 { 00:15:50.664 "name": "BaseBdev1", 00:15:50.664 "uuid": "8f7afaf9-b4ee-4b6b-bc84-7127d8f99045", 00:15:50.664 "is_configured": true, 00:15:50.664 "data_offset": 2048, 00:15:50.664 "data_size": 63488 00:15:50.664 }, 00:15:50.664 { 00:15:50.664 "name": "BaseBdev2", 00:15:50.664 "uuid": "cd3252a8-7e38-4641-b880-26bbe291d8b7", 00:15:50.664 "is_configured": true, 00:15:50.664 "data_offset": 2048, 00:15:50.664 "data_size": 63488 00:15:50.664 }, 00:15:50.664 { 00:15:50.664 "name": "BaseBdev3", 00:15:50.664 "uuid": "57065316-d42e-4947-93b5-56f213c9342f", 00:15:50.664 "is_configured": true, 00:15:50.664 "data_offset": 2048, 00:15:50.664 "data_size": 63488 00:15:50.664 }, 00:15:50.664 { 00:15:50.664 "name": "BaseBdev4", 00:15:50.664 "uuid": "28c942c6-cb9f-4847-8070-e2f428e7c321", 00:15:50.664 "is_configured": true, 00:15:50.664 "data_offset": 2048, 00:15:50.664 "data_size": 63488 00:15:50.664 } 00:15:50.664 ] 00:15:50.664 } 00:15:50.664 } 00:15:50.665 }' 00:15:50.665 15:21:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:50.665 15:21:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:15:50.665 BaseBdev2 00:15:50.665 BaseBdev3 00:15:50.665 BaseBdev4' 00:15:50.665 15:21:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:50.665 15:21:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:15:50.665 15:21:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:50.665 15:21:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:50.665 15:21:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:15:50.665 15:21:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:50.665 15:21:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:50.925 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:50.925 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:50.925 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:50.925 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:50.925 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:15:50.925 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:50.925 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:50.925 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:50.925 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:50.925 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:50.925 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:50.925 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:50.925 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:50.925 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:15:50.925 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:50.925 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:50.925 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:50.925 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:50.925 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:50.925 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:50.925 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:15:50.925 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:50.925 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:50.925 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:50.925 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:50.925 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:50.925 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:50.925 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:15:50.925 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:50.925 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:50.925 [2024-11-19 15:21:41.187740] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:50.925 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:50.925 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:15:50.925 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:15:50.926 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:15:50.926 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:15:50.926 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:15:50.926 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:15:50.926 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:50.926 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:50.926 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:50.926 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:50.926 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:50.926 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:50.926 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:50.926 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:50.926 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:50.926 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:50.926 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:50.926 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:50.926 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:50.926 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:50.926 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:50.926 "name": "Existed_Raid", 00:15:50.926 "uuid": "df0b7e60-da0c-4f86-ad98-f67097e5997e", 00:15:50.926 "strip_size_kb": 64, 00:15:50.926 "state": "online", 00:15:50.926 "raid_level": "raid5f", 00:15:50.926 "superblock": true, 00:15:50.926 "num_base_bdevs": 4, 00:15:50.926 "num_base_bdevs_discovered": 3, 00:15:50.926 "num_base_bdevs_operational": 3, 00:15:50.926 "base_bdevs_list": [ 00:15:50.926 { 00:15:50.926 "name": null, 00:15:50.926 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:50.926 "is_configured": false, 00:15:50.926 "data_offset": 0, 00:15:50.926 "data_size": 63488 00:15:50.926 }, 00:15:50.926 { 00:15:50.926 "name": "BaseBdev2", 00:15:50.926 "uuid": "cd3252a8-7e38-4641-b880-26bbe291d8b7", 00:15:50.926 "is_configured": true, 00:15:50.926 "data_offset": 2048, 00:15:50.926 "data_size": 63488 00:15:50.926 }, 00:15:50.926 { 00:15:50.926 "name": "BaseBdev3", 00:15:50.926 "uuid": "57065316-d42e-4947-93b5-56f213c9342f", 00:15:50.926 "is_configured": true, 00:15:50.926 "data_offset": 2048, 00:15:50.926 "data_size": 63488 00:15:50.926 }, 00:15:50.926 { 00:15:50.926 "name": "BaseBdev4", 00:15:50.926 "uuid": "28c942c6-cb9f-4847-8070-e2f428e7c321", 00:15:50.926 "is_configured": true, 00:15:50.926 "data_offset": 2048, 00:15:50.926 "data_size": 63488 00:15:50.926 } 00:15:50.926 ] 00:15:50.926 }' 00:15:50.926 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:50.926 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.496 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:15:51.496 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:15:51.496 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:51.496 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:51.496 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.496 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:15:51.496 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:51.496 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:15:51.496 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:15:51.496 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:15:51.496 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:51.496 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.496 [2024-11-19 15:21:41.678401] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:15:51.496 [2024-11-19 15:21:41.678605] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:51.496 [2024-11-19 15:21:41.689718] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:51.496 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:51.496 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:15:51.496 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:15:51.496 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:51.496 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:15:51.496 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:51.497 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.497 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:51.497 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:15:51.497 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:15:51.497 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:15:51.497 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:51.497 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.497 [2024-11-19 15:21:41.749650] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:15:51.497 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:51.497 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:15:51.497 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:15:51.497 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:51.497 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:15:51.497 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:51.497 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.497 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:51.497 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:15:51.497 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:15:51.497 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:15:51.497 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:51.497 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.497 [2024-11-19 15:21:41.820678] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:15:51.497 [2024-11-19 15:21:41.820769] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:15:51.497 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:51.497 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:15:51.497 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.758 BaseBdev2 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.758 [ 00:15:51.758 { 00:15:51.758 "name": "BaseBdev2", 00:15:51.758 "aliases": [ 00:15:51.758 "61ba9cf3-f237-4bdb-a7f6-a72620241a91" 00:15:51.758 ], 00:15:51.758 "product_name": "Malloc disk", 00:15:51.758 "block_size": 512, 00:15:51.758 "num_blocks": 65536, 00:15:51.758 "uuid": "61ba9cf3-f237-4bdb-a7f6-a72620241a91", 00:15:51.758 "assigned_rate_limits": { 00:15:51.758 "rw_ios_per_sec": 0, 00:15:51.758 "rw_mbytes_per_sec": 0, 00:15:51.758 "r_mbytes_per_sec": 0, 00:15:51.758 "w_mbytes_per_sec": 0 00:15:51.758 }, 00:15:51.758 "claimed": false, 00:15:51.758 "zoned": false, 00:15:51.758 "supported_io_types": { 00:15:51.758 "read": true, 00:15:51.758 "write": true, 00:15:51.758 "unmap": true, 00:15:51.758 "flush": true, 00:15:51.758 "reset": true, 00:15:51.758 "nvme_admin": false, 00:15:51.758 "nvme_io": false, 00:15:51.758 "nvme_io_md": false, 00:15:51.758 "write_zeroes": true, 00:15:51.758 "zcopy": true, 00:15:51.758 "get_zone_info": false, 00:15:51.758 "zone_management": false, 00:15:51.758 "zone_append": false, 00:15:51.758 "compare": false, 00:15:51.758 "compare_and_write": false, 00:15:51.758 "abort": true, 00:15:51.758 "seek_hole": false, 00:15:51.758 "seek_data": false, 00:15:51.758 "copy": true, 00:15:51.758 "nvme_iov_md": false 00:15:51.758 }, 00:15:51.758 "memory_domains": [ 00:15:51.758 { 00:15:51.758 "dma_device_id": "system", 00:15:51.758 "dma_device_type": 1 00:15:51.758 }, 00:15:51.758 { 00:15:51.758 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:51.758 "dma_device_type": 2 00:15:51.758 } 00:15:51.758 ], 00:15:51.758 "driver_specific": {} 00:15:51.758 } 00:15:51.758 ] 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.758 BaseBdev3 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:51.758 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.758 [ 00:15:51.758 { 00:15:51.758 "name": "BaseBdev3", 00:15:51.758 "aliases": [ 00:15:51.758 "92a63069-250e-44c5-93a1-0fc651741d9a" 00:15:51.758 ], 00:15:51.758 "product_name": "Malloc disk", 00:15:51.758 "block_size": 512, 00:15:51.758 "num_blocks": 65536, 00:15:51.758 "uuid": "92a63069-250e-44c5-93a1-0fc651741d9a", 00:15:51.758 "assigned_rate_limits": { 00:15:51.758 "rw_ios_per_sec": 0, 00:15:51.758 "rw_mbytes_per_sec": 0, 00:15:51.758 "r_mbytes_per_sec": 0, 00:15:51.758 "w_mbytes_per_sec": 0 00:15:51.758 }, 00:15:51.758 "claimed": false, 00:15:51.758 "zoned": false, 00:15:51.758 "supported_io_types": { 00:15:51.759 "read": true, 00:15:51.759 "write": true, 00:15:51.759 "unmap": true, 00:15:51.759 "flush": true, 00:15:51.759 "reset": true, 00:15:51.759 "nvme_admin": false, 00:15:51.759 "nvme_io": false, 00:15:51.759 "nvme_io_md": false, 00:15:51.759 "write_zeroes": true, 00:15:51.759 "zcopy": true, 00:15:51.759 "get_zone_info": false, 00:15:51.759 "zone_management": false, 00:15:51.759 "zone_append": false, 00:15:51.759 "compare": false, 00:15:51.759 "compare_and_write": false, 00:15:51.759 "abort": true, 00:15:51.759 "seek_hole": false, 00:15:51.759 "seek_data": false, 00:15:51.759 "copy": true, 00:15:51.759 "nvme_iov_md": false 00:15:51.759 }, 00:15:51.759 "memory_domains": [ 00:15:51.759 { 00:15:51.759 "dma_device_id": "system", 00:15:51.759 "dma_device_type": 1 00:15:51.759 }, 00:15:51.759 { 00:15:51.759 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:51.759 "dma_device_type": 2 00:15:51.759 } 00:15:51.759 ], 00:15:51.759 "driver_specific": {} 00:15:51.759 } 00:15:51.759 ] 00:15:51.759 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:51.759 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:15:51.759 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:15:51.759 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:15:51.759 15:21:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:15:51.759 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:51.759 15:21:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.759 BaseBdev4 00:15:51.759 15:21:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:51.759 15:21:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:15:51.759 15:21:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:15:51.759 15:21:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:51.759 15:21:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:15:51.759 15:21:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:51.759 15:21:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:51.759 15:21:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:51.759 15:21:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:51.759 15:21:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.759 15:21:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:51.759 15:21:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:15:51.759 15:21:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:51.759 15:21:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.759 [ 00:15:51.759 { 00:15:51.759 "name": "BaseBdev4", 00:15:51.759 "aliases": [ 00:15:51.759 "3bfb6579-00ce-4c7b-8df7-0e15b1c0fbfb" 00:15:51.759 ], 00:15:51.759 "product_name": "Malloc disk", 00:15:51.759 "block_size": 512, 00:15:51.759 "num_blocks": 65536, 00:15:51.759 "uuid": "3bfb6579-00ce-4c7b-8df7-0e15b1c0fbfb", 00:15:51.759 "assigned_rate_limits": { 00:15:51.759 "rw_ios_per_sec": 0, 00:15:51.759 "rw_mbytes_per_sec": 0, 00:15:51.759 "r_mbytes_per_sec": 0, 00:15:51.759 "w_mbytes_per_sec": 0 00:15:51.759 }, 00:15:51.759 "claimed": false, 00:15:51.759 "zoned": false, 00:15:51.759 "supported_io_types": { 00:15:51.759 "read": true, 00:15:51.759 "write": true, 00:15:51.759 "unmap": true, 00:15:51.759 "flush": true, 00:15:51.759 "reset": true, 00:15:51.759 "nvme_admin": false, 00:15:51.759 "nvme_io": false, 00:15:51.759 "nvme_io_md": false, 00:15:51.759 "write_zeroes": true, 00:15:51.759 "zcopy": true, 00:15:51.759 "get_zone_info": false, 00:15:51.759 "zone_management": false, 00:15:51.759 "zone_append": false, 00:15:51.759 "compare": false, 00:15:51.759 "compare_and_write": false, 00:15:51.759 "abort": true, 00:15:51.759 "seek_hole": false, 00:15:51.759 "seek_data": false, 00:15:51.759 "copy": true, 00:15:51.759 "nvme_iov_md": false 00:15:51.759 }, 00:15:51.759 "memory_domains": [ 00:15:51.759 { 00:15:51.759 "dma_device_id": "system", 00:15:51.759 "dma_device_type": 1 00:15:51.759 }, 00:15:51.759 { 00:15:51.759 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:51.759 "dma_device_type": 2 00:15:51.759 } 00:15:51.759 ], 00:15:51.759 "driver_specific": {} 00:15:51.759 } 00:15:51.759 ] 00:15:51.759 15:21:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:51.759 15:21:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:15:51.759 15:21:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:15:51.759 15:21:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:15:51.759 15:21:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:15:51.759 15:21:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:51.759 15:21:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.759 [2024-11-19 15:21:42.052825] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:15:51.759 [2024-11-19 15:21:42.052909] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:15:51.759 [2024-11-19 15:21:42.052956] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:51.759 [2024-11-19 15:21:42.054726] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:15:51.759 [2024-11-19 15:21:42.054809] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:15:51.759 15:21:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:51.759 15:21:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:51.759 15:21:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:51.759 15:21:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:51.759 15:21:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:51.759 15:21:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:51.759 15:21:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:51.759 15:21:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:51.759 15:21:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:51.759 15:21:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:51.759 15:21:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:51.759 15:21:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:51.759 15:21:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:51.759 15:21:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:51.759 15:21:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.759 15:21:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:52.019 15:21:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:52.019 "name": "Existed_Raid", 00:15:52.019 "uuid": "d95a6c03-b8ed-43c2-886f-88a0dd8f7fd6", 00:15:52.019 "strip_size_kb": 64, 00:15:52.019 "state": "configuring", 00:15:52.019 "raid_level": "raid5f", 00:15:52.019 "superblock": true, 00:15:52.019 "num_base_bdevs": 4, 00:15:52.019 "num_base_bdevs_discovered": 3, 00:15:52.019 "num_base_bdevs_operational": 4, 00:15:52.019 "base_bdevs_list": [ 00:15:52.019 { 00:15:52.019 "name": "BaseBdev1", 00:15:52.019 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:52.019 "is_configured": false, 00:15:52.019 "data_offset": 0, 00:15:52.019 "data_size": 0 00:15:52.019 }, 00:15:52.019 { 00:15:52.019 "name": "BaseBdev2", 00:15:52.019 "uuid": "61ba9cf3-f237-4bdb-a7f6-a72620241a91", 00:15:52.019 "is_configured": true, 00:15:52.019 "data_offset": 2048, 00:15:52.019 "data_size": 63488 00:15:52.019 }, 00:15:52.019 { 00:15:52.019 "name": "BaseBdev3", 00:15:52.019 "uuid": "92a63069-250e-44c5-93a1-0fc651741d9a", 00:15:52.019 "is_configured": true, 00:15:52.019 "data_offset": 2048, 00:15:52.020 "data_size": 63488 00:15:52.020 }, 00:15:52.020 { 00:15:52.020 "name": "BaseBdev4", 00:15:52.020 "uuid": "3bfb6579-00ce-4c7b-8df7-0e15b1c0fbfb", 00:15:52.020 "is_configured": true, 00:15:52.020 "data_offset": 2048, 00:15:52.020 "data_size": 63488 00:15:52.020 } 00:15:52.020 ] 00:15:52.020 }' 00:15:52.020 15:21:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:52.020 15:21:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:52.280 15:21:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:15:52.280 15:21:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:52.280 15:21:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:52.280 [2024-11-19 15:21:42.539951] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:15:52.280 15:21:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:52.280 15:21:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:52.280 15:21:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:52.280 15:21:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:52.280 15:21:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:52.280 15:21:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:52.280 15:21:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:52.280 15:21:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:52.280 15:21:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:52.280 15:21:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:52.280 15:21:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:52.280 15:21:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:52.280 15:21:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:52.280 15:21:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:52.280 15:21:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:52.280 15:21:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:52.280 15:21:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:52.280 "name": "Existed_Raid", 00:15:52.280 "uuid": "d95a6c03-b8ed-43c2-886f-88a0dd8f7fd6", 00:15:52.280 "strip_size_kb": 64, 00:15:52.280 "state": "configuring", 00:15:52.280 "raid_level": "raid5f", 00:15:52.280 "superblock": true, 00:15:52.280 "num_base_bdevs": 4, 00:15:52.280 "num_base_bdevs_discovered": 2, 00:15:52.280 "num_base_bdevs_operational": 4, 00:15:52.280 "base_bdevs_list": [ 00:15:52.280 { 00:15:52.280 "name": "BaseBdev1", 00:15:52.280 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:52.280 "is_configured": false, 00:15:52.280 "data_offset": 0, 00:15:52.280 "data_size": 0 00:15:52.280 }, 00:15:52.280 { 00:15:52.280 "name": null, 00:15:52.280 "uuid": "61ba9cf3-f237-4bdb-a7f6-a72620241a91", 00:15:52.280 "is_configured": false, 00:15:52.280 "data_offset": 0, 00:15:52.280 "data_size": 63488 00:15:52.280 }, 00:15:52.280 { 00:15:52.280 "name": "BaseBdev3", 00:15:52.280 "uuid": "92a63069-250e-44c5-93a1-0fc651741d9a", 00:15:52.280 "is_configured": true, 00:15:52.280 "data_offset": 2048, 00:15:52.280 "data_size": 63488 00:15:52.280 }, 00:15:52.280 { 00:15:52.280 "name": "BaseBdev4", 00:15:52.280 "uuid": "3bfb6579-00ce-4c7b-8df7-0e15b1c0fbfb", 00:15:52.280 "is_configured": true, 00:15:52.280 "data_offset": 2048, 00:15:52.280 "data_size": 63488 00:15:52.280 } 00:15:52.280 ] 00:15:52.280 }' 00:15:52.280 15:21:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:52.280 15:21:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:52.850 15:21:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:52.850 15:21:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:52.850 15:21:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:52.850 15:21:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:15:52.850 15:21:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:52.850 15:21:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:15:52.850 15:21:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:15:52.850 15:21:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:52.851 15:21:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:52.851 BaseBdev1 00:15:52.851 [2024-11-19 15:21:43.034137] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:52.851 15:21:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:52.851 15:21:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:15:52.851 15:21:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:15:52.851 15:21:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:52.851 15:21:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:15:52.851 15:21:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:52.851 15:21:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:52.851 15:21:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:52.851 15:21:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:52.851 15:21:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:52.851 15:21:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:52.851 15:21:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:15:52.851 15:21:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:52.851 15:21:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:52.851 [ 00:15:52.851 { 00:15:52.851 "name": "BaseBdev1", 00:15:52.851 "aliases": [ 00:15:52.851 "166523f8-f1bf-49c5-8f8c-4d089768f486" 00:15:52.851 ], 00:15:52.851 "product_name": "Malloc disk", 00:15:52.851 "block_size": 512, 00:15:52.851 "num_blocks": 65536, 00:15:52.851 "uuid": "166523f8-f1bf-49c5-8f8c-4d089768f486", 00:15:52.851 "assigned_rate_limits": { 00:15:52.851 "rw_ios_per_sec": 0, 00:15:52.851 "rw_mbytes_per_sec": 0, 00:15:52.851 "r_mbytes_per_sec": 0, 00:15:52.851 "w_mbytes_per_sec": 0 00:15:52.851 }, 00:15:52.851 "claimed": true, 00:15:52.851 "claim_type": "exclusive_write", 00:15:52.851 "zoned": false, 00:15:52.851 "supported_io_types": { 00:15:52.851 "read": true, 00:15:52.851 "write": true, 00:15:52.851 "unmap": true, 00:15:52.851 "flush": true, 00:15:52.851 "reset": true, 00:15:52.851 "nvme_admin": false, 00:15:52.851 "nvme_io": false, 00:15:52.851 "nvme_io_md": false, 00:15:52.851 "write_zeroes": true, 00:15:52.851 "zcopy": true, 00:15:52.851 "get_zone_info": false, 00:15:52.851 "zone_management": false, 00:15:52.851 "zone_append": false, 00:15:52.851 "compare": false, 00:15:52.851 "compare_and_write": false, 00:15:52.851 "abort": true, 00:15:52.851 "seek_hole": false, 00:15:52.851 "seek_data": false, 00:15:52.851 "copy": true, 00:15:52.851 "nvme_iov_md": false 00:15:52.851 }, 00:15:52.851 "memory_domains": [ 00:15:52.851 { 00:15:52.851 "dma_device_id": "system", 00:15:52.851 "dma_device_type": 1 00:15:52.851 }, 00:15:52.851 { 00:15:52.851 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:52.851 "dma_device_type": 2 00:15:52.851 } 00:15:52.851 ], 00:15:52.851 "driver_specific": {} 00:15:52.851 } 00:15:52.851 ] 00:15:52.851 15:21:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:52.851 15:21:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:15:52.851 15:21:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:52.851 15:21:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:52.851 15:21:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:52.851 15:21:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:52.851 15:21:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:52.851 15:21:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:52.851 15:21:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:52.851 15:21:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:52.851 15:21:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:52.851 15:21:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:52.851 15:21:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:52.851 15:21:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:52.851 15:21:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:52.851 15:21:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:52.851 15:21:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:52.851 15:21:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:52.851 "name": "Existed_Raid", 00:15:52.851 "uuid": "d95a6c03-b8ed-43c2-886f-88a0dd8f7fd6", 00:15:52.851 "strip_size_kb": 64, 00:15:52.851 "state": "configuring", 00:15:52.851 "raid_level": "raid5f", 00:15:52.851 "superblock": true, 00:15:52.851 "num_base_bdevs": 4, 00:15:52.851 "num_base_bdevs_discovered": 3, 00:15:52.851 "num_base_bdevs_operational": 4, 00:15:52.851 "base_bdevs_list": [ 00:15:52.851 { 00:15:52.851 "name": "BaseBdev1", 00:15:52.851 "uuid": "166523f8-f1bf-49c5-8f8c-4d089768f486", 00:15:52.851 "is_configured": true, 00:15:52.851 "data_offset": 2048, 00:15:52.851 "data_size": 63488 00:15:52.851 }, 00:15:52.851 { 00:15:52.851 "name": null, 00:15:52.851 "uuid": "61ba9cf3-f237-4bdb-a7f6-a72620241a91", 00:15:52.851 "is_configured": false, 00:15:52.851 "data_offset": 0, 00:15:52.851 "data_size": 63488 00:15:52.851 }, 00:15:52.851 { 00:15:52.851 "name": "BaseBdev3", 00:15:52.851 "uuid": "92a63069-250e-44c5-93a1-0fc651741d9a", 00:15:52.851 "is_configured": true, 00:15:52.851 "data_offset": 2048, 00:15:52.851 "data_size": 63488 00:15:52.851 }, 00:15:52.851 { 00:15:52.851 "name": "BaseBdev4", 00:15:52.851 "uuid": "3bfb6579-00ce-4c7b-8df7-0e15b1c0fbfb", 00:15:52.851 "is_configured": true, 00:15:52.851 "data_offset": 2048, 00:15:52.851 "data_size": 63488 00:15:52.851 } 00:15:52.851 ] 00:15:52.851 }' 00:15:52.851 15:21:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:52.851 15:21:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:53.422 15:21:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:53.422 15:21:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:53.422 15:21:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:53.422 15:21:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:15:53.422 15:21:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:53.422 15:21:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:15:53.422 15:21:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:15:53.422 15:21:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:53.422 15:21:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:53.422 [2024-11-19 15:21:43.521375] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:15:53.422 15:21:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:53.422 15:21:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:53.422 15:21:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:53.422 15:21:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:53.422 15:21:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:53.422 15:21:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:53.422 15:21:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:53.422 15:21:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:53.422 15:21:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:53.422 15:21:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:53.422 15:21:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:53.422 15:21:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:53.422 15:21:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:53.422 15:21:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:53.422 15:21:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:53.422 15:21:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:53.422 15:21:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:53.422 "name": "Existed_Raid", 00:15:53.422 "uuid": "d95a6c03-b8ed-43c2-886f-88a0dd8f7fd6", 00:15:53.422 "strip_size_kb": 64, 00:15:53.422 "state": "configuring", 00:15:53.422 "raid_level": "raid5f", 00:15:53.422 "superblock": true, 00:15:53.422 "num_base_bdevs": 4, 00:15:53.422 "num_base_bdevs_discovered": 2, 00:15:53.422 "num_base_bdevs_operational": 4, 00:15:53.422 "base_bdevs_list": [ 00:15:53.422 { 00:15:53.422 "name": "BaseBdev1", 00:15:53.422 "uuid": "166523f8-f1bf-49c5-8f8c-4d089768f486", 00:15:53.422 "is_configured": true, 00:15:53.422 "data_offset": 2048, 00:15:53.422 "data_size": 63488 00:15:53.422 }, 00:15:53.422 { 00:15:53.422 "name": null, 00:15:53.422 "uuid": "61ba9cf3-f237-4bdb-a7f6-a72620241a91", 00:15:53.422 "is_configured": false, 00:15:53.422 "data_offset": 0, 00:15:53.422 "data_size": 63488 00:15:53.422 }, 00:15:53.422 { 00:15:53.422 "name": null, 00:15:53.422 "uuid": "92a63069-250e-44c5-93a1-0fc651741d9a", 00:15:53.422 "is_configured": false, 00:15:53.422 "data_offset": 0, 00:15:53.422 "data_size": 63488 00:15:53.422 }, 00:15:53.422 { 00:15:53.422 "name": "BaseBdev4", 00:15:53.422 "uuid": "3bfb6579-00ce-4c7b-8df7-0e15b1c0fbfb", 00:15:53.422 "is_configured": true, 00:15:53.422 "data_offset": 2048, 00:15:53.422 "data_size": 63488 00:15:53.422 } 00:15:53.422 ] 00:15:53.422 }' 00:15:53.422 15:21:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:53.422 15:21:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:53.682 15:21:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:53.682 15:21:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:53.683 15:21:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:15:53.683 15:21:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:53.683 15:21:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:53.683 15:21:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:15:53.683 15:21:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:15:53.683 15:21:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:53.683 15:21:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:53.683 [2024-11-19 15:21:43.996593] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:15:53.683 15:21:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:53.683 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:53.683 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:53.683 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:53.683 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:53.683 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:53.683 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:53.683 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:53.683 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:53.683 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:53.683 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:53.683 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:53.683 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:53.683 15:21:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:53.683 15:21:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:53.943 15:21:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:53.943 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:53.943 "name": "Existed_Raid", 00:15:53.943 "uuid": "d95a6c03-b8ed-43c2-886f-88a0dd8f7fd6", 00:15:53.943 "strip_size_kb": 64, 00:15:53.943 "state": "configuring", 00:15:53.943 "raid_level": "raid5f", 00:15:53.943 "superblock": true, 00:15:53.943 "num_base_bdevs": 4, 00:15:53.943 "num_base_bdevs_discovered": 3, 00:15:53.943 "num_base_bdevs_operational": 4, 00:15:53.943 "base_bdevs_list": [ 00:15:53.943 { 00:15:53.943 "name": "BaseBdev1", 00:15:53.943 "uuid": "166523f8-f1bf-49c5-8f8c-4d089768f486", 00:15:53.943 "is_configured": true, 00:15:53.943 "data_offset": 2048, 00:15:53.943 "data_size": 63488 00:15:53.943 }, 00:15:53.943 { 00:15:53.943 "name": null, 00:15:53.943 "uuid": "61ba9cf3-f237-4bdb-a7f6-a72620241a91", 00:15:53.943 "is_configured": false, 00:15:53.943 "data_offset": 0, 00:15:53.943 "data_size": 63488 00:15:53.943 }, 00:15:53.943 { 00:15:53.943 "name": "BaseBdev3", 00:15:53.943 "uuid": "92a63069-250e-44c5-93a1-0fc651741d9a", 00:15:53.943 "is_configured": true, 00:15:53.943 "data_offset": 2048, 00:15:53.943 "data_size": 63488 00:15:53.943 }, 00:15:53.943 { 00:15:53.943 "name": "BaseBdev4", 00:15:53.943 "uuid": "3bfb6579-00ce-4c7b-8df7-0e15b1c0fbfb", 00:15:53.943 "is_configured": true, 00:15:53.943 "data_offset": 2048, 00:15:53.943 "data_size": 63488 00:15:53.943 } 00:15:53.943 ] 00:15:53.943 }' 00:15:53.943 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:53.943 15:21:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:54.203 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:54.203 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:15:54.203 15:21:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:54.203 15:21:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:54.203 15:21:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:54.203 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:15:54.203 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:15:54.203 15:21:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:54.203 15:21:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:54.203 [2024-11-19 15:21:44.447928] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:54.203 15:21:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:54.203 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:54.203 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:54.203 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:54.203 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:54.203 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:54.203 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:54.203 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:54.203 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:54.203 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:54.203 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:54.203 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:54.203 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:54.203 15:21:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:54.203 15:21:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:54.203 15:21:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:54.203 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:54.203 "name": "Existed_Raid", 00:15:54.203 "uuid": "d95a6c03-b8ed-43c2-886f-88a0dd8f7fd6", 00:15:54.203 "strip_size_kb": 64, 00:15:54.203 "state": "configuring", 00:15:54.203 "raid_level": "raid5f", 00:15:54.203 "superblock": true, 00:15:54.203 "num_base_bdevs": 4, 00:15:54.203 "num_base_bdevs_discovered": 2, 00:15:54.203 "num_base_bdevs_operational": 4, 00:15:54.203 "base_bdevs_list": [ 00:15:54.203 { 00:15:54.203 "name": null, 00:15:54.203 "uuid": "166523f8-f1bf-49c5-8f8c-4d089768f486", 00:15:54.203 "is_configured": false, 00:15:54.203 "data_offset": 0, 00:15:54.203 "data_size": 63488 00:15:54.203 }, 00:15:54.203 { 00:15:54.203 "name": null, 00:15:54.203 "uuid": "61ba9cf3-f237-4bdb-a7f6-a72620241a91", 00:15:54.203 "is_configured": false, 00:15:54.203 "data_offset": 0, 00:15:54.203 "data_size": 63488 00:15:54.203 }, 00:15:54.203 { 00:15:54.203 "name": "BaseBdev3", 00:15:54.203 "uuid": "92a63069-250e-44c5-93a1-0fc651741d9a", 00:15:54.203 "is_configured": true, 00:15:54.203 "data_offset": 2048, 00:15:54.203 "data_size": 63488 00:15:54.203 }, 00:15:54.203 { 00:15:54.203 "name": "BaseBdev4", 00:15:54.203 "uuid": "3bfb6579-00ce-4c7b-8df7-0e15b1c0fbfb", 00:15:54.203 "is_configured": true, 00:15:54.203 "data_offset": 2048, 00:15:54.203 "data_size": 63488 00:15:54.203 } 00:15:54.203 ] 00:15:54.203 }' 00:15:54.203 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:54.203 15:21:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:54.772 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:54.773 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:15:54.773 15:21:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:54.773 15:21:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:54.773 15:21:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:54.773 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:15:54.773 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:15:54.773 15:21:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:54.773 15:21:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:54.773 [2024-11-19 15:21:44.925684] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:54.773 15:21:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:54.773 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:54.773 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:54.773 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:54.773 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:54.773 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:54.773 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:54.773 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:54.773 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:54.773 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:54.773 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:54.773 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:54.773 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:54.773 15:21:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:54.773 15:21:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:54.773 15:21:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:54.773 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:54.773 "name": "Existed_Raid", 00:15:54.773 "uuid": "d95a6c03-b8ed-43c2-886f-88a0dd8f7fd6", 00:15:54.773 "strip_size_kb": 64, 00:15:54.773 "state": "configuring", 00:15:54.773 "raid_level": "raid5f", 00:15:54.773 "superblock": true, 00:15:54.773 "num_base_bdevs": 4, 00:15:54.773 "num_base_bdevs_discovered": 3, 00:15:54.773 "num_base_bdevs_operational": 4, 00:15:54.773 "base_bdevs_list": [ 00:15:54.773 { 00:15:54.773 "name": null, 00:15:54.773 "uuid": "166523f8-f1bf-49c5-8f8c-4d089768f486", 00:15:54.773 "is_configured": false, 00:15:54.773 "data_offset": 0, 00:15:54.773 "data_size": 63488 00:15:54.773 }, 00:15:54.773 { 00:15:54.773 "name": "BaseBdev2", 00:15:54.773 "uuid": "61ba9cf3-f237-4bdb-a7f6-a72620241a91", 00:15:54.773 "is_configured": true, 00:15:54.773 "data_offset": 2048, 00:15:54.773 "data_size": 63488 00:15:54.773 }, 00:15:54.773 { 00:15:54.773 "name": "BaseBdev3", 00:15:54.773 "uuid": "92a63069-250e-44c5-93a1-0fc651741d9a", 00:15:54.773 "is_configured": true, 00:15:54.773 "data_offset": 2048, 00:15:54.773 "data_size": 63488 00:15:54.773 }, 00:15:54.773 { 00:15:54.773 "name": "BaseBdev4", 00:15:54.773 "uuid": "3bfb6579-00ce-4c7b-8df7-0e15b1c0fbfb", 00:15:54.773 "is_configured": true, 00:15:54.773 "data_offset": 2048, 00:15:54.773 "data_size": 63488 00:15:54.773 } 00:15:54.773 ] 00:15:54.773 }' 00:15:54.773 15:21:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:54.773 15:21:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 166523f8-f1bf-49c5-8f8c-4d089768f486 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:55.344 [2024-11-19 15:21:45.475529] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:15:55.344 [2024-11-19 15:21:45.475695] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:15:55.344 [2024-11-19 15:21:45.475708] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:15:55.344 [2024-11-19 15:21:45.475987] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:15:55.344 NewBaseBdev 00:15:55.344 [2024-11-19 15:21:45.476465] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:15:55.344 [2024-11-19 15:21:45.476484] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:15:55.344 [2024-11-19 15:21:45.476584] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:55.344 [ 00:15:55.344 { 00:15:55.344 "name": "NewBaseBdev", 00:15:55.344 "aliases": [ 00:15:55.344 "166523f8-f1bf-49c5-8f8c-4d089768f486" 00:15:55.344 ], 00:15:55.344 "product_name": "Malloc disk", 00:15:55.344 "block_size": 512, 00:15:55.344 "num_blocks": 65536, 00:15:55.344 "uuid": "166523f8-f1bf-49c5-8f8c-4d089768f486", 00:15:55.344 "assigned_rate_limits": { 00:15:55.344 "rw_ios_per_sec": 0, 00:15:55.344 "rw_mbytes_per_sec": 0, 00:15:55.344 "r_mbytes_per_sec": 0, 00:15:55.344 "w_mbytes_per_sec": 0 00:15:55.344 }, 00:15:55.344 "claimed": true, 00:15:55.344 "claim_type": "exclusive_write", 00:15:55.344 "zoned": false, 00:15:55.344 "supported_io_types": { 00:15:55.344 "read": true, 00:15:55.344 "write": true, 00:15:55.344 "unmap": true, 00:15:55.344 "flush": true, 00:15:55.344 "reset": true, 00:15:55.344 "nvme_admin": false, 00:15:55.344 "nvme_io": false, 00:15:55.344 "nvme_io_md": false, 00:15:55.344 "write_zeroes": true, 00:15:55.344 "zcopy": true, 00:15:55.344 "get_zone_info": false, 00:15:55.344 "zone_management": false, 00:15:55.344 "zone_append": false, 00:15:55.344 "compare": false, 00:15:55.344 "compare_and_write": false, 00:15:55.344 "abort": true, 00:15:55.344 "seek_hole": false, 00:15:55.344 "seek_data": false, 00:15:55.344 "copy": true, 00:15:55.344 "nvme_iov_md": false 00:15:55.344 }, 00:15:55.344 "memory_domains": [ 00:15:55.344 { 00:15:55.344 "dma_device_id": "system", 00:15:55.344 "dma_device_type": 1 00:15:55.344 }, 00:15:55.344 { 00:15:55.344 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:55.344 "dma_device_type": 2 00:15:55.344 } 00:15:55.344 ], 00:15:55.344 "driver_specific": {} 00:15:55.344 } 00:15:55.344 ] 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:55.344 15:21:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:55.344 "name": "Existed_Raid", 00:15:55.344 "uuid": "d95a6c03-b8ed-43c2-886f-88a0dd8f7fd6", 00:15:55.344 "strip_size_kb": 64, 00:15:55.344 "state": "online", 00:15:55.344 "raid_level": "raid5f", 00:15:55.344 "superblock": true, 00:15:55.344 "num_base_bdevs": 4, 00:15:55.344 "num_base_bdevs_discovered": 4, 00:15:55.344 "num_base_bdevs_operational": 4, 00:15:55.344 "base_bdevs_list": [ 00:15:55.344 { 00:15:55.344 "name": "NewBaseBdev", 00:15:55.344 "uuid": "166523f8-f1bf-49c5-8f8c-4d089768f486", 00:15:55.344 "is_configured": true, 00:15:55.344 "data_offset": 2048, 00:15:55.344 "data_size": 63488 00:15:55.344 }, 00:15:55.344 { 00:15:55.344 "name": "BaseBdev2", 00:15:55.344 "uuid": "61ba9cf3-f237-4bdb-a7f6-a72620241a91", 00:15:55.344 "is_configured": true, 00:15:55.344 "data_offset": 2048, 00:15:55.344 "data_size": 63488 00:15:55.344 }, 00:15:55.344 { 00:15:55.344 "name": "BaseBdev3", 00:15:55.344 "uuid": "92a63069-250e-44c5-93a1-0fc651741d9a", 00:15:55.345 "is_configured": true, 00:15:55.345 "data_offset": 2048, 00:15:55.345 "data_size": 63488 00:15:55.345 }, 00:15:55.345 { 00:15:55.345 "name": "BaseBdev4", 00:15:55.345 "uuid": "3bfb6579-00ce-4c7b-8df7-0e15b1c0fbfb", 00:15:55.345 "is_configured": true, 00:15:55.345 "data_offset": 2048, 00:15:55.345 "data_size": 63488 00:15:55.345 } 00:15:55.345 ] 00:15:55.345 }' 00:15:55.345 15:21:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:55.345 15:21:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:55.914 15:21:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:15:55.914 15:21:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:15:55.914 15:21:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:55.914 15:21:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:55.914 15:21:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:15:55.915 15:21:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:55.915 15:21:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:15:55.915 15:21:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:55.915 15:21:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:55.915 15:21:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:55.915 [2024-11-19 15:21:45.954927] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:55.915 15:21:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:55.915 15:21:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:55.915 "name": "Existed_Raid", 00:15:55.915 "aliases": [ 00:15:55.915 "d95a6c03-b8ed-43c2-886f-88a0dd8f7fd6" 00:15:55.915 ], 00:15:55.915 "product_name": "Raid Volume", 00:15:55.915 "block_size": 512, 00:15:55.915 "num_blocks": 190464, 00:15:55.915 "uuid": "d95a6c03-b8ed-43c2-886f-88a0dd8f7fd6", 00:15:55.915 "assigned_rate_limits": { 00:15:55.915 "rw_ios_per_sec": 0, 00:15:55.915 "rw_mbytes_per_sec": 0, 00:15:55.915 "r_mbytes_per_sec": 0, 00:15:55.915 "w_mbytes_per_sec": 0 00:15:55.915 }, 00:15:55.915 "claimed": false, 00:15:55.915 "zoned": false, 00:15:55.915 "supported_io_types": { 00:15:55.915 "read": true, 00:15:55.915 "write": true, 00:15:55.915 "unmap": false, 00:15:55.915 "flush": false, 00:15:55.915 "reset": true, 00:15:55.915 "nvme_admin": false, 00:15:55.915 "nvme_io": false, 00:15:55.915 "nvme_io_md": false, 00:15:55.915 "write_zeroes": true, 00:15:55.915 "zcopy": false, 00:15:55.915 "get_zone_info": false, 00:15:55.915 "zone_management": false, 00:15:55.915 "zone_append": false, 00:15:55.915 "compare": false, 00:15:55.915 "compare_and_write": false, 00:15:55.915 "abort": false, 00:15:55.915 "seek_hole": false, 00:15:55.915 "seek_data": false, 00:15:55.915 "copy": false, 00:15:55.915 "nvme_iov_md": false 00:15:55.915 }, 00:15:55.915 "driver_specific": { 00:15:55.915 "raid": { 00:15:55.915 "uuid": "d95a6c03-b8ed-43c2-886f-88a0dd8f7fd6", 00:15:55.915 "strip_size_kb": 64, 00:15:55.915 "state": "online", 00:15:55.915 "raid_level": "raid5f", 00:15:55.915 "superblock": true, 00:15:55.915 "num_base_bdevs": 4, 00:15:55.915 "num_base_bdevs_discovered": 4, 00:15:55.915 "num_base_bdevs_operational": 4, 00:15:55.915 "base_bdevs_list": [ 00:15:55.915 { 00:15:55.915 "name": "NewBaseBdev", 00:15:55.915 "uuid": "166523f8-f1bf-49c5-8f8c-4d089768f486", 00:15:55.915 "is_configured": true, 00:15:55.915 "data_offset": 2048, 00:15:55.915 "data_size": 63488 00:15:55.915 }, 00:15:55.915 { 00:15:55.915 "name": "BaseBdev2", 00:15:55.915 "uuid": "61ba9cf3-f237-4bdb-a7f6-a72620241a91", 00:15:55.915 "is_configured": true, 00:15:55.915 "data_offset": 2048, 00:15:55.915 "data_size": 63488 00:15:55.915 }, 00:15:55.915 { 00:15:55.915 "name": "BaseBdev3", 00:15:55.915 "uuid": "92a63069-250e-44c5-93a1-0fc651741d9a", 00:15:55.915 "is_configured": true, 00:15:55.915 "data_offset": 2048, 00:15:55.915 "data_size": 63488 00:15:55.915 }, 00:15:55.915 { 00:15:55.915 "name": "BaseBdev4", 00:15:55.915 "uuid": "3bfb6579-00ce-4c7b-8df7-0e15b1c0fbfb", 00:15:55.915 "is_configured": true, 00:15:55.915 "data_offset": 2048, 00:15:55.915 "data_size": 63488 00:15:55.915 } 00:15:55.915 ] 00:15:55.915 } 00:15:55.915 } 00:15:55.915 }' 00:15:55.915 15:21:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:55.915 15:21:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:15:55.915 BaseBdev2 00:15:55.915 BaseBdev3 00:15:55.915 BaseBdev4' 00:15:55.915 15:21:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:55.915 15:21:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:15:55.915 15:21:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:55.915 15:21:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:15:55.915 15:21:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:55.915 15:21:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:55.915 15:21:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:55.915 15:21:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:55.915 15:21:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:55.915 15:21:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:55.915 15:21:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:55.915 15:21:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:15:55.915 15:21:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:55.915 15:21:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:55.915 15:21:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:55.915 15:21:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:55.915 15:21:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:55.915 15:21:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:55.915 15:21:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:55.915 15:21:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:55.915 15:21:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:15:55.915 15:21:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:55.915 15:21:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:55.915 15:21:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:55.915 15:21:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:55.915 15:21:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:55.915 15:21:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:55.915 15:21:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:55.915 15:21:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:15:55.915 15:21:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:55.915 15:21:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:55.915 15:21:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:56.175 15:21:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:56.175 15:21:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:56.175 15:21:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:15:56.175 15:21:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:56.175 15:21:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:56.175 [2024-11-19 15:21:46.258225] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:15:56.175 [2024-11-19 15:21:46.258253] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:56.175 [2024-11-19 15:21:46.258315] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:56.176 [2024-11-19 15:21:46.258576] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:56.176 [2024-11-19 15:21:46.258594] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:15:56.176 15:21:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:56.176 15:21:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 93907 00:15:56.176 15:21:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 93907 ']' 00:15:56.176 15:21:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 93907 00:15:56.176 15:21:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:15:56.176 15:21:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:15:56.176 15:21:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 93907 00:15:56.176 killing process with pid 93907 00:15:56.176 15:21:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:15:56.176 15:21:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:15:56.176 15:21:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 93907' 00:15:56.176 15:21:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 93907 00:15:56.176 [2024-11-19 15:21:46.293072] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:56.176 15:21:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 93907 00:15:56.176 [2024-11-19 15:21:46.334104] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:56.436 ************************************ 00:15:56.436 END TEST raid5f_state_function_test_sb 00:15:56.436 15:21:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:15:56.436 00:15:56.436 real 0m9.489s 00:15:56.436 user 0m16.195s 00:15:56.436 sys 0m2.141s 00:15:56.436 15:21:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:15:56.436 15:21:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:56.436 ************************************ 00:15:56.436 15:21:46 bdev_raid -- bdev/bdev_raid.sh@988 -- # run_test raid5f_superblock_test raid_superblock_test raid5f 4 00:15:56.436 15:21:46 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:15:56.436 15:21:46 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:15:56.436 15:21:46 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:56.436 ************************************ 00:15:56.436 START TEST raid5f_superblock_test 00:15:56.437 ************************************ 00:15:56.437 15:21:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid5f 4 00:15:56.437 15:21:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid5f 00:15:56.437 15:21:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:15:56.437 15:21:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:15:56.437 15:21:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:15:56.437 15:21:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:15:56.437 15:21:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:15:56.437 15:21:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:15:56.437 15:21:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:15:56.437 15:21:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:15:56.437 15:21:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:15:56.437 15:21:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:15:56.437 15:21:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:15:56.437 15:21:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:15:56.437 15:21:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid5f '!=' raid1 ']' 00:15:56.437 15:21:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:15:56.437 15:21:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:15:56.437 15:21:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=94555 00:15:56.437 15:21:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:15:56.437 15:21:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 94555 00:15:56.437 15:21:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 94555 ']' 00:15:56.437 15:21:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:56.437 15:21:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:15:56.437 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:56.437 15:21:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:56.437 15:21:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:15:56.437 15:21:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:56.437 [2024-11-19 15:21:46.711541] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:15:56.437 [2024-11-19 15:21:46.711674] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid94555 ] 00:15:56.697 [2024-11-19 15:21:46.867042] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:56.697 [2024-11-19 15:21:46.891067] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:56.697 [2024-11-19 15:21:46.934186] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:56.697 [2024-11-19 15:21:46.934227] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:57.268 15:21:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:15:57.268 15:21:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:15:57.268 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:15:57.268 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:57.268 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:15:57.268 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:15:57.268 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:15:57.268 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:15:57.268 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:15:57.268 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:15:57.268 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:15:57.268 15:21:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:57.268 15:21:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:57.268 malloc1 00:15:57.268 15:21:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:57.268 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:57.268 15:21:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:57.268 15:21:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:57.268 [2024-11-19 15:21:47.572788] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:57.268 [2024-11-19 15:21:47.572848] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:57.268 [2024-11-19 15:21:47.572872] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:15:57.268 [2024-11-19 15:21:47.572886] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:57.268 [2024-11-19 15:21:47.574959] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:57.268 [2024-11-19 15:21:47.575010] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:57.268 pt1 00:15:57.268 15:21:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:57.268 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:15:57.268 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:57.268 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:15:57.268 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:15:57.268 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:15:57.268 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:15:57.268 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:15:57.268 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:15:57.268 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:15:57.268 15:21:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:57.268 15:21:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:57.268 malloc2 00:15:57.268 15:21:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:57.268 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:57.268 15:21:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:57.268 15:21:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:57.268 [2024-11-19 15:21:47.601526] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:57.268 [2024-11-19 15:21:47.601579] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:57.268 [2024-11-19 15:21:47.601593] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:15:57.268 [2024-11-19 15:21:47.601604] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:57.268 [2024-11-19 15:21:47.603629] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:57.268 [2024-11-19 15:21:47.603666] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:57.529 pt2 00:15:57.529 15:21:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:57.529 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:15:57.529 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:57.529 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:15:57.529 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:15:57.529 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:15:57.529 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:15:57.529 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:15:57.529 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:15:57.529 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:15:57.529 15:21:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:57.529 15:21:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:57.529 malloc3 00:15:57.529 15:21:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:57.529 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:15:57.529 15:21:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:57.529 15:21:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:57.529 [2024-11-19 15:21:47.630313] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:15:57.529 [2024-11-19 15:21:47.630367] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:57.529 [2024-11-19 15:21:47.630385] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:15:57.529 [2024-11-19 15:21:47.630394] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:57.529 [2024-11-19 15:21:47.632413] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:57.529 [2024-11-19 15:21:47.632453] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:15:57.529 pt3 00:15:57.529 15:21:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:57.529 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:15:57.529 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:57.529 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:15:57.529 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:15:57.529 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:15:57.529 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:15:57.529 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:15:57.529 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:15:57.529 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:15:57.529 15:21:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:57.529 15:21:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:57.529 malloc4 00:15:57.529 15:21:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:57.529 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:15:57.529 15:21:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:57.529 15:21:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:57.529 [2024-11-19 15:21:47.680386] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:15:57.529 [2024-11-19 15:21:47.680461] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:57.529 [2024-11-19 15:21:47.680485] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:15:57.529 [2024-11-19 15:21:47.680504] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:57.529 [2024-11-19 15:21:47.683791] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:57.529 [2024-11-19 15:21:47.683845] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:15:57.529 pt4 00:15:57.529 15:21:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:57.529 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:15:57.529 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:57.529 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:15:57.529 15:21:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:57.529 15:21:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:57.530 [2024-11-19 15:21:47.692499] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:57.530 [2024-11-19 15:21:47.694631] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:57.530 [2024-11-19 15:21:47.694710] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:15:57.530 [2024-11-19 15:21:47.694763] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:15:57.530 [2024-11-19 15:21:47.694948] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:15:57.530 [2024-11-19 15:21:47.694983] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:15:57.530 [2024-11-19 15:21:47.695256] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:15:57.530 [2024-11-19 15:21:47.695800] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:15:57.530 [2024-11-19 15:21:47.695827] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:15:57.530 [2024-11-19 15:21:47.695998] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:57.530 15:21:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:57.530 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:15:57.530 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:57.530 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:57.530 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:57.530 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:57.530 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:57.530 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:57.530 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:57.530 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:57.530 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:57.530 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:57.530 15:21:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:57.530 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:57.530 15:21:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:57.530 15:21:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:57.530 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:57.530 "name": "raid_bdev1", 00:15:57.530 "uuid": "80dc3004-3449-4ce9-b903-d5b1a3c22d20", 00:15:57.530 "strip_size_kb": 64, 00:15:57.530 "state": "online", 00:15:57.530 "raid_level": "raid5f", 00:15:57.530 "superblock": true, 00:15:57.530 "num_base_bdevs": 4, 00:15:57.530 "num_base_bdevs_discovered": 4, 00:15:57.530 "num_base_bdevs_operational": 4, 00:15:57.530 "base_bdevs_list": [ 00:15:57.530 { 00:15:57.530 "name": "pt1", 00:15:57.530 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:57.530 "is_configured": true, 00:15:57.530 "data_offset": 2048, 00:15:57.530 "data_size": 63488 00:15:57.530 }, 00:15:57.530 { 00:15:57.530 "name": "pt2", 00:15:57.530 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:57.530 "is_configured": true, 00:15:57.530 "data_offset": 2048, 00:15:57.530 "data_size": 63488 00:15:57.530 }, 00:15:57.530 { 00:15:57.530 "name": "pt3", 00:15:57.530 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:57.530 "is_configured": true, 00:15:57.530 "data_offset": 2048, 00:15:57.530 "data_size": 63488 00:15:57.530 }, 00:15:57.530 { 00:15:57.530 "name": "pt4", 00:15:57.530 "uuid": "00000000-0000-0000-0000-000000000004", 00:15:57.530 "is_configured": true, 00:15:57.530 "data_offset": 2048, 00:15:57.530 "data_size": 63488 00:15:57.530 } 00:15:57.530 ] 00:15:57.530 }' 00:15:57.530 15:21:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:57.530 15:21:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:57.791 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:15:57.791 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:15:57.791 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:57.791 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:57.791 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:15:57.791 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:57.791 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:57.791 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:57.791 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:57.791 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:57.791 [2024-11-19 15:21:48.109326] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:58.051 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:58.051 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:58.051 "name": "raid_bdev1", 00:15:58.051 "aliases": [ 00:15:58.051 "80dc3004-3449-4ce9-b903-d5b1a3c22d20" 00:15:58.051 ], 00:15:58.051 "product_name": "Raid Volume", 00:15:58.051 "block_size": 512, 00:15:58.051 "num_blocks": 190464, 00:15:58.051 "uuid": "80dc3004-3449-4ce9-b903-d5b1a3c22d20", 00:15:58.051 "assigned_rate_limits": { 00:15:58.051 "rw_ios_per_sec": 0, 00:15:58.051 "rw_mbytes_per_sec": 0, 00:15:58.051 "r_mbytes_per_sec": 0, 00:15:58.051 "w_mbytes_per_sec": 0 00:15:58.051 }, 00:15:58.051 "claimed": false, 00:15:58.051 "zoned": false, 00:15:58.051 "supported_io_types": { 00:15:58.051 "read": true, 00:15:58.051 "write": true, 00:15:58.051 "unmap": false, 00:15:58.051 "flush": false, 00:15:58.051 "reset": true, 00:15:58.051 "nvme_admin": false, 00:15:58.051 "nvme_io": false, 00:15:58.051 "nvme_io_md": false, 00:15:58.051 "write_zeroes": true, 00:15:58.051 "zcopy": false, 00:15:58.051 "get_zone_info": false, 00:15:58.051 "zone_management": false, 00:15:58.051 "zone_append": false, 00:15:58.051 "compare": false, 00:15:58.051 "compare_and_write": false, 00:15:58.051 "abort": false, 00:15:58.051 "seek_hole": false, 00:15:58.051 "seek_data": false, 00:15:58.051 "copy": false, 00:15:58.051 "nvme_iov_md": false 00:15:58.051 }, 00:15:58.051 "driver_specific": { 00:15:58.051 "raid": { 00:15:58.051 "uuid": "80dc3004-3449-4ce9-b903-d5b1a3c22d20", 00:15:58.051 "strip_size_kb": 64, 00:15:58.051 "state": "online", 00:15:58.051 "raid_level": "raid5f", 00:15:58.051 "superblock": true, 00:15:58.051 "num_base_bdevs": 4, 00:15:58.051 "num_base_bdevs_discovered": 4, 00:15:58.051 "num_base_bdevs_operational": 4, 00:15:58.051 "base_bdevs_list": [ 00:15:58.051 { 00:15:58.051 "name": "pt1", 00:15:58.051 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:58.051 "is_configured": true, 00:15:58.051 "data_offset": 2048, 00:15:58.051 "data_size": 63488 00:15:58.051 }, 00:15:58.051 { 00:15:58.051 "name": "pt2", 00:15:58.051 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:58.051 "is_configured": true, 00:15:58.051 "data_offset": 2048, 00:15:58.051 "data_size": 63488 00:15:58.051 }, 00:15:58.051 { 00:15:58.051 "name": "pt3", 00:15:58.051 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:58.051 "is_configured": true, 00:15:58.051 "data_offset": 2048, 00:15:58.051 "data_size": 63488 00:15:58.051 }, 00:15:58.051 { 00:15:58.051 "name": "pt4", 00:15:58.051 "uuid": "00000000-0000-0000-0000-000000000004", 00:15:58.051 "is_configured": true, 00:15:58.051 "data_offset": 2048, 00:15:58.051 "data_size": 63488 00:15:58.051 } 00:15:58.051 ] 00:15:58.051 } 00:15:58.051 } 00:15:58.051 }' 00:15:58.051 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:58.051 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:15:58.051 pt2 00:15:58.051 pt3 00:15:58.051 pt4' 00:15:58.051 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:58.051 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:15:58.051 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:58.051 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:58.051 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:15:58.051 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:58.051 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:58.052 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:58.052 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:58.052 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:58.052 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:58.052 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:58.052 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:15:58.052 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:58.052 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:58.052 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:58.052 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:58.052 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:58.052 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:58.052 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:15:58.052 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:58.052 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:58.052 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:58.052 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:58.052 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:58.052 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:58.052 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:58.052 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:15:58.052 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:58.052 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:58.052 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:58.052 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:58.052 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:58.052 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:58.312 [2024-11-19 15:21:48.396826] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=80dc3004-3449-4ce9-b903-d5b1a3c22d20 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 80dc3004-3449-4ce9-b903-d5b1a3c22d20 ']' 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:58.312 [2024-11-19 15:21:48.424646] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:58.312 [2024-11-19 15:21:48.424677] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:58.312 [2024-11-19 15:21:48.424733] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:58.312 [2024-11-19 15:21:48.424806] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:58.312 [2024-11-19 15:21:48.424821] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:58.312 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:58.313 [2024-11-19 15:21:48.588374] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:15:58.313 [2024-11-19 15:21:48.590186] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:15:58.313 [2024-11-19 15:21:48.590246] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:15:58.313 [2024-11-19 15:21:48.590273] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:15:58.313 [2024-11-19 15:21:48.590311] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:15:58.313 [2024-11-19 15:21:48.590345] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:15:58.313 [2024-11-19 15:21:48.590364] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:15:58.313 [2024-11-19 15:21:48.590379] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:15:58.313 [2024-11-19 15:21:48.590391] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:58.313 [2024-11-19 15:21:48.590409] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:15:58.313 request: 00:15:58.313 { 00:15:58.313 "name": "raid_bdev1", 00:15:58.313 "raid_level": "raid5f", 00:15:58.313 "base_bdevs": [ 00:15:58.313 "malloc1", 00:15:58.313 "malloc2", 00:15:58.313 "malloc3", 00:15:58.313 "malloc4" 00:15:58.313 ], 00:15:58.313 "strip_size_kb": 64, 00:15:58.313 "superblock": false, 00:15:58.313 "method": "bdev_raid_create", 00:15:58.313 "req_id": 1 00:15:58.313 } 00:15:58.313 Got JSON-RPC error response 00:15:58.313 response: 00:15:58.313 { 00:15:58.313 "code": -17, 00:15:58.313 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:15:58.313 } 00:15:58.313 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:15:58.313 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:15:58.313 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:15:58.313 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:15:58.313 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:15:58.313 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:58.313 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:15:58.313 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:58.313 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:58.313 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:58.313 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:15:58.313 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:15:58.313 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:58.313 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:58.313 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:58.573 [2024-11-19 15:21:48.652234] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:58.573 [2024-11-19 15:21:48.652279] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:58.573 [2024-11-19 15:21:48.652296] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:15:58.573 [2024-11-19 15:21:48.652304] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:58.573 [2024-11-19 15:21:48.654326] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:58.573 [2024-11-19 15:21:48.654361] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:58.573 [2024-11-19 15:21:48.654417] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:15:58.573 [2024-11-19 15:21:48.654447] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:58.573 pt1 00:15:58.573 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:58.573 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 4 00:15:58.573 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:58.573 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:58.573 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:58.573 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:58.573 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:58.573 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:58.573 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:58.573 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:58.573 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:58.573 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:58.573 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:58.573 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:58.573 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:58.573 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:58.573 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:58.573 "name": "raid_bdev1", 00:15:58.573 "uuid": "80dc3004-3449-4ce9-b903-d5b1a3c22d20", 00:15:58.573 "strip_size_kb": 64, 00:15:58.573 "state": "configuring", 00:15:58.573 "raid_level": "raid5f", 00:15:58.573 "superblock": true, 00:15:58.573 "num_base_bdevs": 4, 00:15:58.573 "num_base_bdevs_discovered": 1, 00:15:58.573 "num_base_bdevs_operational": 4, 00:15:58.573 "base_bdevs_list": [ 00:15:58.573 { 00:15:58.573 "name": "pt1", 00:15:58.573 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:58.573 "is_configured": true, 00:15:58.573 "data_offset": 2048, 00:15:58.573 "data_size": 63488 00:15:58.573 }, 00:15:58.573 { 00:15:58.573 "name": null, 00:15:58.573 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:58.573 "is_configured": false, 00:15:58.573 "data_offset": 2048, 00:15:58.573 "data_size": 63488 00:15:58.573 }, 00:15:58.573 { 00:15:58.573 "name": null, 00:15:58.573 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:58.573 "is_configured": false, 00:15:58.573 "data_offset": 2048, 00:15:58.573 "data_size": 63488 00:15:58.573 }, 00:15:58.573 { 00:15:58.573 "name": null, 00:15:58.573 "uuid": "00000000-0000-0000-0000-000000000004", 00:15:58.573 "is_configured": false, 00:15:58.573 "data_offset": 2048, 00:15:58.573 "data_size": 63488 00:15:58.573 } 00:15:58.573 ] 00:15:58.573 }' 00:15:58.573 15:21:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:58.573 15:21:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:58.834 15:21:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:15:58.834 15:21:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:58.834 15:21:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:58.834 15:21:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:58.834 [2024-11-19 15:21:49.111480] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:58.834 [2024-11-19 15:21:49.111530] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:58.834 [2024-11-19 15:21:49.111548] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:15:58.834 [2024-11-19 15:21:49.111576] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:58.834 [2024-11-19 15:21:49.111929] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:58.834 [2024-11-19 15:21:49.111954] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:58.834 [2024-11-19 15:21:49.112031] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:15:58.834 [2024-11-19 15:21:49.112054] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:58.834 pt2 00:15:58.834 15:21:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:58.834 15:21:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:15:58.834 15:21:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:58.834 15:21:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:58.834 [2024-11-19 15:21:49.123483] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:15:58.834 15:21:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:58.834 15:21:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 4 00:15:58.834 15:21:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:58.834 15:21:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:58.834 15:21:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:58.834 15:21:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:58.834 15:21:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:58.834 15:21:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:58.834 15:21:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:58.834 15:21:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:58.834 15:21:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:58.834 15:21:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:58.834 15:21:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:58.834 15:21:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:58.834 15:21:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:58.834 15:21:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:59.094 15:21:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:59.094 "name": "raid_bdev1", 00:15:59.094 "uuid": "80dc3004-3449-4ce9-b903-d5b1a3c22d20", 00:15:59.094 "strip_size_kb": 64, 00:15:59.094 "state": "configuring", 00:15:59.094 "raid_level": "raid5f", 00:15:59.094 "superblock": true, 00:15:59.094 "num_base_bdevs": 4, 00:15:59.094 "num_base_bdevs_discovered": 1, 00:15:59.094 "num_base_bdevs_operational": 4, 00:15:59.094 "base_bdevs_list": [ 00:15:59.094 { 00:15:59.094 "name": "pt1", 00:15:59.094 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:59.094 "is_configured": true, 00:15:59.094 "data_offset": 2048, 00:15:59.094 "data_size": 63488 00:15:59.094 }, 00:15:59.094 { 00:15:59.094 "name": null, 00:15:59.094 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:59.094 "is_configured": false, 00:15:59.094 "data_offset": 0, 00:15:59.094 "data_size": 63488 00:15:59.094 }, 00:15:59.094 { 00:15:59.094 "name": null, 00:15:59.094 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:59.094 "is_configured": false, 00:15:59.094 "data_offset": 2048, 00:15:59.094 "data_size": 63488 00:15:59.094 }, 00:15:59.094 { 00:15:59.094 "name": null, 00:15:59.094 "uuid": "00000000-0000-0000-0000-000000000004", 00:15:59.094 "is_configured": false, 00:15:59.094 "data_offset": 2048, 00:15:59.094 "data_size": 63488 00:15:59.094 } 00:15:59.094 ] 00:15:59.094 }' 00:15:59.094 15:21:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:59.094 15:21:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:59.354 15:21:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:15:59.354 15:21:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:15:59.354 15:21:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:59.354 15:21:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:59.354 15:21:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:59.354 [2024-11-19 15:21:49.566731] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:59.354 [2024-11-19 15:21:49.566805] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:59.354 [2024-11-19 15:21:49.566823] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:15:59.354 [2024-11-19 15:21:49.566834] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:59.354 [2024-11-19 15:21:49.567133] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:59.354 [2024-11-19 15:21:49.567175] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:59.354 [2024-11-19 15:21:49.567227] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:15:59.354 [2024-11-19 15:21:49.567246] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:59.354 pt2 00:15:59.354 15:21:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:59.354 15:21:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:15:59.354 15:21:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:15:59.354 15:21:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:15:59.354 15:21:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:59.354 15:21:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:59.354 [2024-11-19 15:21:49.578683] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:15:59.354 [2024-11-19 15:21:49.578733] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:59.354 [2024-11-19 15:21:49.578747] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:15:59.354 [2024-11-19 15:21:49.578757] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:59.354 [2024-11-19 15:21:49.579098] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:59.354 [2024-11-19 15:21:49.579125] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:15:59.354 [2024-11-19 15:21:49.579173] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:15:59.354 [2024-11-19 15:21:49.579191] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:15:59.354 pt3 00:15:59.354 15:21:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:59.354 15:21:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:15:59.354 15:21:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:15:59.354 15:21:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:15:59.354 15:21:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:59.354 15:21:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:59.354 [2024-11-19 15:21:49.590667] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:15:59.354 [2024-11-19 15:21:49.590718] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:59.354 [2024-11-19 15:21:49.590731] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:15:59.354 [2024-11-19 15:21:49.590740] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:59.354 [2024-11-19 15:21:49.591029] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:59.354 [2024-11-19 15:21:49.591055] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:15:59.354 [2024-11-19 15:21:49.591101] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:15:59.354 [2024-11-19 15:21:49.591120] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:15:59.354 [2024-11-19 15:21:49.591208] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:15:59.354 [2024-11-19 15:21:49.591219] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:15:59.354 [2024-11-19 15:21:49.591431] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:15:59.354 [2024-11-19 15:21:49.591912] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:15:59.354 [2024-11-19 15:21:49.591933] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:15:59.354 [2024-11-19 15:21:49.592037] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:59.354 pt4 00:15:59.354 15:21:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:59.354 15:21:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:15:59.354 15:21:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:15:59.354 15:21:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:15:59.354 15:21:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:59.354 15:21:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:59.354 15:21:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:59.354 15:21:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:59.354 15:21:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:59.354 15:21:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:59.354 15:21:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:59.354 15:21:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:59.354 15:21:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:59.354 15:21:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:59.354 15:21:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:59.354 15:21:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:59.354 15:21:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:59.355 15:21:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:59.355 15:21:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:59.355 "name": "raid_bdev1", 00:15:59.355 "uuid": "80dc3004-3449-4ce9-b903-d5b1a3c22d20", 00:15:59.355 "strip_size_kb": 64, 00:15:59.355 "state": "online", 00:15:59.355 "raid_level": "raid5f", 00:15:59.355 "superblock": true, 00:15:59.355 "num_base_bdevs": 4, 00:15:59.355 "num_base_bdevs_discovered": 4, 00:15:59.355 "num_base_bdevs_operational": 4, 00:15:59.355 "base_bdevs_list": [ 00:15:59.355 { 00:15:59.355 "name": "pt1", 00:15:59.355 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:59.355 "is_configured": true, 00:15:59.355 "data_offset": 2048, 00:15:59.355 "data_size": 63488 00:15:59.355 }, 00:15:59.355 { 00:15:59.355 "name": "pt2", 00:15:59.355 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:59.355 "is_configured": true, 00:15:59.355 "data_offset": 2048, 00:15:59.355 "data_size": 63488 00:15:59.355 }, 00:15:59.355 { 00:15:59.355 "name": "pt3", 00:15:59.355 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:59.355 "is_configured": true, 00:15:59.355 "data_offset": 2048, 00:15:59.355 "data_size": 63488 00:15:59.355 }, 00:15:59.355 { 00:15:59.355 "name": "pt4", 00:15:59.355 "uuid": "00000000-0000-0000-0000-000000000004", 00:15:59.355 "is_configured": true, 00:15:59.355 "data_offset": 2048, 00:15:59.355 "data_size": 63488 00:15:59.355 } 00:15:59.355 ] 00:15:59.355 }' 00:15:59.355 15:21:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:59.355 15:21:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:59.927 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:15:59.927 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:15:59.927 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:59.927 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:59.927 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:15:59.928 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:59.928 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:59.928 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:59.928 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:59.928 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:59.928 [2024-11-19 15:21:50.078043] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:59.928 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:59.928 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:59.928 "name": "raid_bdev1", 00:15:59.928 "aliases": [ 00:15:59.928 "80dc3004-3449-4ce9-b903-d5b1a3c22d20" 00:15:59.928 ], 00:15:59.928 "product_name": "Raid Volume", 00:15:59.928 "block_size": 512, 00:15:59.928 "num_blocks": 190464, 00:15:59.928 "uuid": "80dc3004-3449-4ce9-b903-d5b1a3c22d20", 00:15:59.928 "assigned_rate_limits": { 00:15:59.928 "rw_ios_per_sec": 0, 00:15:59.928 "rw_mbytes_per_sec": 0, 00:15:59.928 "r_mbytes_per_sec": 0, 00:15:59.928 "w_mbytes_per_sec": 0 00:15:59.928 }, 00:15:59.928 "claimed": false, 00:15:59.928 "zoned": false, 00:15:59.928 "supported_io_types": { 00:15:59.928 "read": true, 00:15:59.928 "write": true, 00:15:59.928 "unmap": false, 00:15:59.928 "flush": false, 00:15:59.928 "reset": true, 00:15:59.928 "nvme_admin": false, 00:15:59.928 "nvme_io": false, 00:15:59.928 "nvme_io_md": false, 00:15:59.928 "write_zeroes": true, 00:15:59.928 "zcopy": false, 00:15:59.928 "get_zone_info": false, 00:15:59.928 "zone_management": false, 00:15:59.928 "zone_append": false, 00:15:59.928 "compare": false, 00:15:59.928 "compare_and_write": false, 00:15:59.928 "abort": false, 00:15:59.928 "seek_hole": false, 00:15:59.929 "seek_data": false, 00:15:59.929 "copy": false, 00:15:59.929 "nvme_iov_md": false 00:15:59.929 }, 00:15:59.929 "driver_specific": { 00:15:59.929 "raid": { 00:15:59.929 "uuid": "80dc3004-3449-4ce9-b903-d5b1a3c22d20", 00:15:59.929 "strip_size_kb": 64, 00:15:59.929 "state": "online", 00:15:59.929 "raid_level": "raid5f", 00:15:59.929 "superblock": true, 00:15:59.929 "num_base_bdevs": 4, 00:15:59.929 "num_base_bdevs_discovered": 4, 00:15:59.929 "num_base_bdevs_operational": 4, 00:15:59.929 "base_bdevs_list": [ 00:15:59.929 { 00:15:59.929 "name": "pt1", 00:15:59.929 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:59.929 "is_configured": true, 00:15:59.929 "data_offset": 2048, 00:15:59.929 "data_size": 63488 00:15:59.929 }, 00:15:59.929 { 00:15:59.929 "name": "pt2", 00:15:59.929 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:59.929 "is_configured": true, 00:15:59.929 "data_offset": 2048, 00:15:59.929 "data_size": 63488 00:15:59.929 }, 00:15:59.929 { 00:15:59.929 "name": "pt3", 00:15:59.929 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:59.929 "is_configured": true, 00:15:59.929 "data_offset": 2048, 00:15:59.929 "data_size": 63488 00:15:59.929 }, 00:15:59.929 { 00:15:59.929 "name": "pt4", 00:15:59.929 "uuid": "00000000-0000-0000-0000-000000000004", 00:15:59.929 "is_configured": true, 00:15:59.929 "data_offset": 2048, 00:15:59.929 "data_size": 63488 00:15:59.929 } 00:15:59.929 ] 00:15:59.929 } 00:15:59.929 } 00:15:59.929 }' 00:15:59.930 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:59.930 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:15:59.930 pt2 00:15:59.930 pt3 00:15:59.930 pt4' 00:15:59.930 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:59.930 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:15:59.930 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:59.930 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:15:59.930 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:59.930 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:59.930 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:59.930 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:59.930 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:59.930 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:59.930 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:59.930 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:59.930 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:15:59.930 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:59.930 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:59.930 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:59.930 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:59.930 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:59.930 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:00.192 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:16:00.192 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:00.192 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:00.192 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:00.193 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:00.193 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:16:00.193 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:16:00.193 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:00.193 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:00.193 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:16:00.193 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:00.193 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:00.193 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:00.193 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:16:00.193 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:16:00.193 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:16:00.193 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:00.193 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:00.193 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:00.193 [2024-11-19 15:21:50.353521] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:00.193 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:00.193 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 80dc3004-3449-4ce9-b903-d5b1a3c22d20 '!=' 80dc3004-3449-4ce9-b903-d5b1a3c22d20 ']' 00:16:00.193 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid5f 00:16:00.193 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:16:00.193 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:16:00.193 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:16:00.193 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:00.193 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:00.193 [2024-11-19 15:21:50.381351] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:16:00.193 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:00.193 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:16:00.193 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:00.193 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:00.193 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:16:00.193 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:16:00.193 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:16:00.193 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:00.193 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:00.193 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:00.193 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:00.193 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:00.193 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:00.193 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:00.193 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:00.193 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:00.193 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:00.193 "name": "raid_bdev1", 00:16:00.193 "uuid": "80dc3004-3449-4ce9-b903-d5b1a3c22d20", 00:16:00.193 "strip_size_kb": 64, 00:16:00.193 "state": "online", 00:16:00.193 "raid_level": "raid5f", 00:16:00.193 "superblock": true, 00:16:00.193 "num_base_bdevs": 4, 00:16:00.193 "num_base_bdevs_discovered": 3, 00:16:00.193 "num_base_bdevs_operational": 3, 00:16:00.193 "base_bdevs_list": [ 00:16:00.193 { 00:16:00.193 "name": null, 00:16:00.193 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:00.193 "is_configured": false, 00:16:00.193 "data_offset": 0, 00:16:00.193 "data_size": 63488 00:16:00.193 }, 00:16:00.193 { 00:16:00.193 "name": "pt2", 00:16:00.193 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:00.193 "is_configured": true, 00:16:00.193 "data_offset": 2048, 00:16:00.193 "data_size": 63488 00:16:00.193 }, 00:16:00.193 { 00:16:00.193 "name": "pt3", 00:16:00.193 "uuid": "00000000-0000-0000-0000-000000000003", 00:16:00.193 "is_configured": true, 00:16:00.193 "data_offset": 2048, 00:16:00.193 "data_size": 63488 00:16:00.193 }, 00:16:00.193 { 00:16:00.193 "name": "pt4", 00:16:00.193 "uuid": "00000000-0000-0000-0000-000000000004", 00:16:00.193 "is_configured": true, 00:16:00.193 "data_offset": 2048, 00:16:00.193 "data_size": 63488 00:16:00.193 } 00:16:00.193 ] 00:16:00.193 }' 00:16:00.193 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:00.193 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:00.763 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:00.763 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:00.763 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:00.763 [2024-11-19 15:21:50.840521] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:00.763 [2024-11-19 15:21:50.840546] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:00.763 [2024-11-19 15:21:50.840600] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:00.763 [2024-11-19 15:21:50.840655] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:00.763 [2024-11-19 15:21:50.840666] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:16:00.763 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:00.763 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:00.763 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:00.763 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:16:00.763 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:00.763 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:00.763 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:16:00.763 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:16:00.763 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:16:00.763 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:16:00.763 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:16:00.763 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:00.763 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:00.763 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:00.763 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:16:00.763 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:16:00.763 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:16:00.763 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:00.763 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:00.763 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:00.763 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:16:00.763 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:16:00.763 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt4 00:16:00.763 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:00.763 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:00.763 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:00.763 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:16:00.763 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:16:00.763 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:16:00.763 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:16:00.764 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:00.764 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:00.764 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:00.764 [2024-11-19 15:21:50.940347] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:00.764 [2024-11-19 15:21:50.940395] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:00.764 [2024-11-19 15:21:50.940408] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:16:00.764 [2024-11-19 15:21:50.940418] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:00.764 [2024-11-19 15:21:50.942620] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:00.764 [2024-11-19 15:21:50.942658] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:00.764 [2024-11-19 15:21:50.942715] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:16:00.764 [2024-11-19 15:21:50.942762] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:00.764 pt2 00:16:00.764 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:00.764 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:16:00.764 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:00.764 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:00.764 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:16:00.764 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:16:00.764 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:16:00.764 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:00.764 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:00.764 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:00.764 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:00.764 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:00.764 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:00.764 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:00.764 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:00.764 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:00.764 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:00.764 "name": "raid_bdev1", 00:16:00.764 "uuid": "80dc3004-3449-4ce9-b903-d5b1a3c22d20", 00:16:00.764 "strip_size_kb": 64, 00:16:00.764 "state": "configuring", 00:16:00.764 "raid_level": "raid5f", 00:16:00.764 "superblock": true, 00:16:00.764 "num_base_bdevs": 4, 00:16:00.764 "num_base_bdevs_discovered": 1, 00:16:00.764 "num_base_bdevs_operational": 3, 00:16:00.764 "base_bdevs_list": [ 00:16:00.764 { 00:16:00.764 "name": null, 00:16:00.764 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:00.764 "is_configured": false, 00:16:00.764 "data_offset": 2048, 00:16:00.764 "data_size": 63488 00:16:00.764 }, 00:16:00.764 { 00:16:00.764 "name": "pt2", 00:16:00.764 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:00.764 "is_configured": true, 00:16:00.764 "data_offset": 2048, 00:16:00.764 "data_size": 63488 00:16:00.764 }, 00:16:00.764 { 00:16:00.764 "name": null, 00:16:00.764 "uuid": "00000000-0000-0000-0000-000000000003", 00:16:00.764 "is_configured": false, 00:16:00.764 "data_offset": 2048, 00:16:00.764 "data_size": 63488 00:16:00.764 }, 00:16:00.764 { 00:16:00.764 "name": null, 00:16:00.764 "uuid": "00000000-0000-0000-0000-000000000004", 00:16:00.764 "is_configured": false, 00:16:00.764 "data_offset": 2048, 00:16:00.764 "data_size": 63488 00:16:00.764 } 00:16:00.764 ] 00:16:00.764 }' 00:16:00.764 15:21:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:00.764 15:21:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:01.334 15:21:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:16:01.334 15:21:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:16:01.334 15:21:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:16:01.334 15:21:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:01.334 15:21:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:01.334 [2024-11-19 15:21:51.411623] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:16:01.334 [2024-11-19 15:21:51.411680] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:01.334 [2024-11-19 15:21:51.411696] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:16:01.334 [2024-11-19 15:21:51.411707] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:01.334 [2024-11-19 15:21:51.412063] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:01.334 [2024-11-19 15:21:51.412090] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:16:01.334 [2024-11-19 15:21:51.412146] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:16:01.334 [2024-11-19 15:21:51.412168] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:16:01.334 pt3 00:16:01.334 15:21:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:01.334 15:21:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:16:01.334 15:21:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:01.334 15:21:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:01.334 15:21:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:16:01.334 15:21:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:16:01.334 15:21:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:16:01.334 15:21:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:01.334 15:21:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:01.334 15:21:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:01.334 15:21:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:01.334 15:21:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:01.334 15:21:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:01.334 15:21:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:01.334 15:21:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:01.334 15:21:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:01.334 15:21:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:01.334 "name": "raid_bdev1", 00:16:01.334 "uuid": "80dc3004-3449-4ce9-b903-d5b1a3c22d20", 00:16:01.334 "strip_size_kb": 64, 00:16:01.334 "state": "configuring", 00:16:01.334 "raid_level": "raid5f", 00:16:01.334 "superblock": true, 00:16:01.334 "num_base_bdevs": 4, 00:16:01.334 "num_base_bdevs_discovered": 2, 00:16:01.334 "num_base_bdevs_operational": 3, 00:16:01.334 "base_bdevs_list": [ 00:16:01.334 { 00:16:01.334 "name": null, 00:16:01.334 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:01.334 "is_configured": false, 00:16:01.334 "data_offset": 2048, 00:16:01.334 "data_size": 63488 00:16:01.334 }, 00:16:01.334 { 00:16:01.334 "name": "pt2", 00:16:01.334 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:01.334 "is_configured": true, 00:16:01.334 "data_offset": 2048, 00:16:01.334 "data_size": 63488 00:16:01.334 }, 00:16:01.334 { 00:16:01.334 "name": "pt3", 00:16:01.334 "uuid": "00000000-0000-0000-0000-000000000003", 00:16:01.334 "is_configured": true, 00:16:01.334 "data_offset": 2048, 00:16:01.334 "data_size": 63488 00:16:01.334 }, 00:16:01.334 { 00:16:01.334 "name": null, 00:16:01.334 "uuid": "00000000-0000-0000-0000-000000000004", 00:16:01.334 "is_configured": false, 00:16:01.334 "data_offset": 2048, 00:16:01.334 "data_size": 63488 00:16:01.334 } 00:16:01.334 ] 00:16:01.334 }' 00:16:01.334 15:21:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:01.334 15:21:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:01.595 15:21:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:16:01.595 15:21:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:16:01.595 15:21:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@519 -- # i=3 00:16:01.595 15:21:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:16:01.595 15:21:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:01.595 15:21:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:01.595 [2024-11-19 15:21:51.874837] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:16:01.595 [2024-11-19 15:21:51.874899] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:01.595 [2024-11-19 15:21:51.874915] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:16:01.595 [2024-11-19 15:21:51.874925] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:01.595 [2024-11-19 15:21:51.875247] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:01.595 [2024-11-19 15:21:51.875274] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:16:01.595 [2024-11-19 15:21:51.875324] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:16:01.595 [2024-11-19 15:21:51.875342] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:16:01.595 [2024-11-19 15:21:51.875422] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:16:01.595 [2024-11-19 15:21:51.875434] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:16:01.595 [2024-11-19 15:21:51.875653] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:16:01.595 [2024-11-19 15:21:51.876208] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:16:01.595 [2024-11-19 15:21:51.876229] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:16:01.595 [2024-11-19 15:21:51.876446] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:01.595 pt4 00:16:01.595 15:21:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:01.595 15:21:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:16:01.595 15:21:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:01.595 15:21:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:01.595 15:21:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:16:01.595 15:21:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:16:01.595 15:21:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:16:01.595 15:21:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:01.595 15:21:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:01.595 15:21:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:01.595 15:21:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:01.595 15:21:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:01.595 15:21:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:01.595 15:21:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:01.595 15:21:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:01.595 15:21:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:01.595 15:21:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:01.595 "name": "raid_bdev1", 00:16:01.595 "uuid": "80dc3004-3449-4ce9-b903-d5b1a3c22d20", 00:16:01.595 "strip_size_kb": 64, 00:16:01.595 "state": "online", 00:16:01.595 "raid_level": "raid5f", 00:16:01.595 "superblock": true, 00:16:01.595 "num_base_bdevs": 4, 00:16:01.595 "num_base_bdevs_discovered": 3, 00:16:01.595 "num_base_bdevs_operational": 3, 00:16:01.595 "base_bdevs_list": [ 00:16:01.595 { 00:16:01.595 "name": null, 00:16:01.595 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:01.595 "is_configured": false, 00:16:01.595 "data_offset": 2048, 00:16:01.595 "data_size": 63488 00:16:01.595 }, 00:16:01.595 { 00:16:01.595 "name": "pt2", 00:16:01.595 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:01.595 "is_configured": true, 00:16:01.595 "data_offset": 2048, 00:16:01.595 "data_size": 63488 00:16:01.595 }, 00:16:01.595 { 00:16:01.595 "name": "pt3", 00:16:01.595 "uuid": "00000000-0000-0000-0000-000000000003", 00:16:01.595 "is_configured": true, 00:16:01.595 "data_offset": 2048, 00:16:01.595 "data_size": 63488 00:16:01.595 }, 00:16:01.596 { 00:16:01.596 "name": "pt4", 00:16:01.596 "uuid": "00000000-0000-0000-0000-000000000004", 00:16:01.596 "is_configured": true, 00:16:01.596 "data_offset": 2048, 00:16:01.596 "data_size": 63488 00:16:01.596 } 00:16:01.596 ] 00:16:01.596 }' 00:16:01.596 15:21:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:01.856 15:21:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:02.116 15:21:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:02.116 15:21:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:02.116 15:21:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:02.116 [2024-11-19 15:21:52.314080] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:02.116 [2024-11-19 15:21:52.314109] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:02.116 [2024-11-19 15:21:52.314185] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:02.116 [2024-11-19 15:21:52.314271] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:02.116 [2024-11-19 15:21:52.314300] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:16:02.116 15:21:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:02.116 15:21:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:02.116 15:21:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:16:02.116 15:21:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:02.116 15:21:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:02.116 15:21:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:02.116 15:21:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:16:02.116 15:21:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:16:02.116 15:21:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 4 -gt 2 ']' 00:16:02.116 15:21:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@534 -- # i=3 00:16:02.116 15:21:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt4 00:16:02.116 15:21:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:02.116 15:21:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:02.116 15:21:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:02.116 15:21:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:02.116 15:21:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:02.116 15:21:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:02.116 [2024-11-19 15:21:52.386041] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:02.116 [2024-11-19 15:21:52.386086] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:02.116 [2024-11-19 15:21:52.386102] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b480 00:16:02.116 [2024-11-19 15:21:52.386110] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:02.116 [2024-11-19 15:21:52.388341] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:02.116 [2024-11-19 15:21:52.388380] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:02.116 [2024-11-19 15:21:52.388441] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:16:02.116 [2024-11-19 15:21:52.388471] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:02.116 [2024-11-19 15:21:52.388570] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:16:02.116 [2024-11-19 15:21:52.388610] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:02.116 [2024-11-19 15:21:52.388632] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:16:02.116 [2024-11-19 15:21:52.388671] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:02.116 [2024-11-19 15:21:52.388776] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:16:02.116 pt1 00:16:02.116 15:21:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:02.116 15:21:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 4 -gt 2 ']' 00:16:02.116 15:21:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:16:02.116 15:21:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:02.116 15:21:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:02.116 15:21:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:16:02.116 15:21:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:16:02.116 15:21:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:16:02.116 15:21:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:02.116 15:21:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:02.116 15:21:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:02.116 15:21:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:02.116 15:21:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:02.116 15:21:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:02.116 15:21:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:02.116 15:21:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:02.116 15:21:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:02.116 15:21:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:02.116 "name": "raid_bdev1", 00:16:02.116 "uuid": "80dc3004-3449-4ce9-b903-d5b1a3c22d20", 00:16:02.116 "strip_size_kb": 64, 00:16:02.116 "state": "configuring", 00:16:02.116 "raid_level": "raid5f", 00:16:02.116 "superblock": true, 00:16:02.116 "num_base_bdevs": 4, 00:16:02.116 "num_base_bdevs_discovered": 2, 00:16:02.116 "num_base_bdevs_operational": 3, 00:16:02.116 "base_bdevs_list": [ 00:16:02.116 { 00:16:02.116 "name": null, 00:16:02.116 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:02.116 "is_configured": false, 00:16:02.116 "data_offset": 2048, 00:16:02.116 "data_size": 63488 00:16:02.116 }, 00:16:02.116 { 00:16:02.116 "name": "pt2", 00:16:02.116 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:02.116 "is_configured": true, 00:16:02.116 "data_offset": 2048, 00:16:02.116 "data_size": 63488 00:16:02.116 }, 00:16:02.116 { 00:16:02.116 "name": "pt3", 00:16:02.116 "uuid": "00000000-0000-0000-0000-000000000003", 00:16:02.116 "is_configured": true, 00:16:02.116 "data_offset": 2048, 00:16:02.116 "data_size": 63488 00:16:02.116 }, 00:16:02.116 { 00:16:02.116 "name": null, 00:16:02.116 "uuid": "00000000-0000-0000-0000-000000000004", 00:16:02.116 "is_configured": false, 00:16:02.116 "data_offset": 2048, 00:16:02.116 "data_size": 63488 00:16:02.116 } 00:16:02.116 ] 00:16:02.116 }' 00:16:02.116 15:21:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:02.116 15:21:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:02.686 15:21:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:16:02.686 15:21:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:02.686 15:21:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:02.686 15:21:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:16:02.686 15:21:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:02.686 15:21:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:16:02.686 15:21:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:16:02.686 15:21:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:02.686 15:21:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:02.686 [2024-11-19 15:21:52.897124] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:16:02.686 [2024-11-19 15:21:52.897173] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:02.686 [2024-11-19 15:21:52.897187] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ba80 00:16:02.686 [2024-11-19 15:21:52.897197] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:02.686 [2024-11-19 15:21:52.897530] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:02.686 [2024-11-19 15:21:52.897559] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:16:02.686 [2024-11-19 15:21:52.897613] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:16:02.686 [2024-11-19 15:21:52.897642] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:16:02.686 [2024-11-19 15:21:52.897734] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:16:02.686 [2024-11-19 15:21:52.897745] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:16:02.686 [2024-11-19 15:21:52.897959] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:16:02.686 [2024-11-19 15:21:52.898504] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:16:02.686 [2024-11-19 15:21:52.898531] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:16:02.686 [2024-11-19 15:21:52.898704] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:02.686 pt4 00:16:02.686 15:21:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:02.686 15:21:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:16:02.686 15:21:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:02.686 15:21:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:02.686 15:21:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:16:02.686 15:21:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:16:02.686 15:21:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:16:02.686 15:21:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:02.686 15:21:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:02.686 15:21:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:02.686 15:21:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:02.686 15:21:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:02.686 15:21:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:02.686 15:21:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:02.687 15:21:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:02.687 15:21:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:02.687 15:21:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:02.687 "name": "raid_bdev1", 00:16:02.687 "uuid": "80dc3004-3449-4ce9-b903-d5b1a3c22d20", 00:16:02.687 "strip_size_kb": 64, 00:16:02.687 "state": "online", 00:16:02.687 "raid_level": "raid5f", 00:16:02.687 "superblock": true, 00:16:02.687 "num_base_bdevs": 4, 00:16:02.687 "num_base_bdevs_discovered": 3, 00:16:02.687 "num_base_bdevs_operational": 3, 00:16:02.687 "base_bdevs_list": [ 00:16:02.687 { 00:16:02.687 "name": null, 00:16:02.687 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:02.687 "is_configured": false, 00:16:02.687 "data_offset": 2048, 00:16:02.687 "data_size": 63488 00:16:02.687 }, 00:16:02.687 { 00:16:02.687 "name": "pt2", 00:16:02.687 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:02.687 "is_configured": true, 00:16:02.687 "data_offset": 2048, 00:16:02.687 "data_size": 63488 00:16:02.687 }, 00:16:02.687 { 00:16:02.687 "name": "pt3", 00:16:02.687 "uuid": "00000000-0000-0000-0000-000000000003", 00:16:02.687 "is_configured": true, 00:16:02.687 "data_offset": 2048, 00:16:02.687 "data_size": 63488 00:16:02.687 }, 00:16:02.687 { 00:16:02.687 "name": "pt4", 00:16:02.687 "uuid": "00000000-0000-0000-0000-000000000004", 00:16:02.687 "is_configured": true, 00:16:02.687 "data_offset": 2048, 00:16:02.687 "data_size": 63488 00:16:02.687 } 00:16:02.687 ] 00:16:02.687 }' 00:16:02.687 15:21:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:02.687 15:21:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:03.272 15:21:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:16:03.272 15:21:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:16:03.272 15:21:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:03.272 15:21:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:03.272 15:21:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:03.272 15:21:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:16:03.272 15:21:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:16:03.273 15:21:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:03.273 15:21:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:03.273 15:21:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:03.273 [2024-11-19 15:21:53.404427] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:03.273 15:21:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:03.273 15:21:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' 80dc3004-3449-4ce9-b903-d5b1a3c22d20 '!=' 80dc3004-3449-4ce9-b903-d5b1a3c22d20 ']' 00:16:03.273 15:21:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 94555 00:16:03.273 15:21:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 94555 ']' 00:16:03.273 15:21:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@958 -- # kill -0 94555 00:16:03.273 15:21:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@959 -- # uname 00:16:03.273 15:21:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:16:03.273 15:21:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 94555 00:16:03.273 15:21:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:16:03.273 15:21:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:16:03.273 killing process with pid 94555 00:16:03.273 15:21:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 94555' 00:16:03.273 15:21:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@973 -- # kill 94555 00:16:03.273 [2024-11-19 15:21:53.470667] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:03.273 [2024-11-19 15:21:53.470745] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:03.273 [2024-11-19 15:21:53.470821] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:03.273 [2024-11-19 15:21:53.470835] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:16:03.273 15:21:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@978 -- # wait 94555 00:16:03.273 [2024-11-19 15:21:53.514932] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:03.558 15:21:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:16:03.558 00:16:03.558 real 0m7.111s 00:16:03.558 user 0m11.979s 00:16:03.558 sys 0m1.570s 00:16:03.558 15:21:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:03.558 15:21:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:16:03.558 ************************************ 00:16:03.558 END TEST raid5f_superblock_test 00:16:03.558 ************************************ 00:16:03.558 15:21:53 bdev_raid -- bdev/bdev_raid.sh@989 -- # '[' true = true ']' 00:16:03.558 15:21:53 bdev_raid -- bdev/bdev_raid.sh@990 -- # run_test raid5f_rebuild_test raid_rebuild_test raid5f 4 false false true 00:16:03.558 15:21:53 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:16:03.559 15:21:53 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:03.559 15:21:53 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:03.559 ************************************ 00:16:03.559 START TEST raid5f_rebuild_test 00:16:03.559 ************************************ 00:16:03.559 15:21:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid5f 4 false false true 00:16:03.559 15:21:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:16:03.559 15:21:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:16:03.559 15:21:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:16:03.559 15:21:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:16:03.559 15:21:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:16:03.559 15:21:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:16:03.559 15:21:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:03.559 15:21:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:16:03.559 15:21:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:16:03.559 15:21:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:03.559 15:21:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:16:03.559 15:21:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:16:03.559 15:21:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:03.559 15:21:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:16:03.559 15:21:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:16:03.559 15:21:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:03.559 15:21:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:16:03.559 15:21:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:16:03.559 15:21:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:03.559 15:21:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:16:03.559 15:21:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:16:03.559 15:21:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:16:03.559 15:21:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:16:03.559 15:21:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:16:03.559 15:21:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:16:03.559 15:21:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:16:03.559 15:21:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:16:03.559 15:21:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:16:03.559 15:21:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:16:03.559 15:21:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:16:03.559 15:21:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:16:03.559 15:21:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=95024 00:16:03.559 15:21:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:16:03.559 15:21:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 95024 00:16:03.559 15:21:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@835 -- # '[' -z 95024 ']' 00:16:03.559 15:21:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:03.559 15:21:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:16:03.559 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:03.559 15:21:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:03.559 15:21:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:16:03.559 15:21:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:03.834 I/O size of 3145728 is greater than zero copy threshold (65536). 00:16:03.834 Zero copy mechanism will not be used. 00:16:03.834 [2024-11-19 15:21:53.911188] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:16:03.834 [2024-11-19 15:21:53.911317] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid95024 ] 00:16:03.834 [2024-11-19 15:21:54.057214] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:03.834 [2024-11-19 15:21:54.081269] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:03.834 [2024-11-19 15:21:54.124250] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:03.834 [2024-11-19 15:21:54.124285] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:04.403 15:21:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:16:04.403 15:21:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@868 -- # return 0 00:16:04.403 15:21:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:16:04.403 15:21:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:16:04.403 15:21:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:04.403 15:21:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:04.664 BaseBdev1_malloc 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:04.664 [2024-11-19 15:21:54.754535] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:16:04.664 [2024-11-19 15:21:54.754594] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:04.664 [2024-11-19 15:21:54.754625] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:16:04.664 [2024-11-19 15:21:54.754644] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:04.664 [2024-11-19 15:21:54.756690] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:04.664 [2024-11-19 15:21:54.756725] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:16:04.664 BaseBdev1 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:04.664 BaseBdev2_malloc 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:04.664 [2024-11-19 15:21:54.783038] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:16:04.664 [2024-11-19 15:21:54.783098] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:04.664 [2024-11-19 15:21:54.783117] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:16:04.664 [2024-11-19 15:21:54.783125] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:04.664 [2024-11-19 15:21:54.785131] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:04.664 [2024-11-19 15:21:54.785173] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:16:04.664 BaseBdev2 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:04.664 BaseBdev3_malloc 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:04.664 [2024-11-19 15:21:54.811510] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:16:04.664 [2024-11-19 15:21:54.811561] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:04.664 [2024-11-19 15:21:54.811581] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:16:04.664 [2024-11-19 15:21:54.811589] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:04.664 [2024-11-19 15:21:54.813587] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:04.664 [2024-11-19 15:21:54.813622] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:16:04.664 BaseBdev3 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:04.664 BaseBdev4_malloc 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:04.664 [2024-11-19 15:21:54.857821] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:16:04.664 [2024-11-19 15:21:54.857910] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:04.664 [2024-11-19 15:21:54.857953] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:16:04.664 [2024-11-19 15:21:54.858000] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:04.664 [2024-11-19 15:21:54.862154] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:04.664 [2024-11-19 15:21:54.862216] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:16:04.664 BaseBdev4 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:04.664 spare_malloc 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:04.664 spare_delay 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:04.664 [2024-11-19 15:21:54.899725] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:04.664 [2024-11-19 15:21:54.899771] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:04.664 [2024-11-19 15:21:54.899795] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:16:04.664 [2024-11-19 15:21:54.899804] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:04.664 [2024-11-19 15:21:54.902009] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:04.664 [2024-11-19 15:21:54.902039] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:04.664 spare 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:04.664 [2024-11-19 15:21:54.911775] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:04.664 [2024-11-19 15:21:54.913529] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:04.664 [2024-11-19 15:21:54.913590] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:16:04.664 [2024-11-19 15:21:54.913635] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:16:04.664 [2024-11-19 15:21:54.913733] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:16:04.664 [2024-11-19 15:21:54.913753] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:16:04.664 [2024-11-19 15:21:54.913994] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:16:04.664 [2024-11-19 15:21:54.914411] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:16:04.664 [2024-11-19 15:21:54.914431] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:16:04.664 [2024-11-19 15:21:54.914555] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:16:04.664 15:21:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:16:04.665 15:21:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:04.665 15:21:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:04.665 15:21:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:04.665 15:21:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:04.665 15:21:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:04.665 15:21:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:04.665 15:21:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:04.665 15:21:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:04.665 15:21:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:04.665 15:21:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:04.665 "name": "raid_bdev1", 00:16:04.665 "uuid": "9539b9e4-8c91-4dcd-8499-03e3c8ef07ea", 00:16:04.665 "strip_size_kb": 64, 00:16:04.665 "state": "online", 00:16:04.665 "raid_level": "raid5f", 00:16:04.665 "superblock": false, 00:16:04.665 "num_base_bdevs": 4, 00:16:04.665 "num_base_bdevs_discovered": 4, 00:16:04.665 "num_base_bdevs_operational": 4, 00:16:04.665 "base_bdevs_list": [ 00:16:04.665 { 00:16:04.665 "name": "BaseBdev1", 00:16:04.665 "uuid": "7fa0ae05-ca60-5bb6-b036-c869c402dd0d", 00:16:04.665 "is_configured": true, 00:16:04.665 "data_offset": 0, 00:16:04.665 "data_size": 65536 00:16:04.665 }, 00:16:04.665 { 00:16:04.665 "name": "BaseBdev2", 00:16:04.665 "uuid": "06197ab8-31a3-55c0-893d-cbd85e6e3e17", 00:16:04.665 "is_configured": true, 00:16:04.665 "data_offset": 0, 00:16:04.665 "data_size": 65536 00:16:04.665 }, 00:16:04.665 { 00:16:04.665 "name": "BaseBdev3", 00:16:04.665 "uuid": "0163c5e4-ac32-5fd6-a6cc-5260e8058920", 00:16:04.665 "is_configured": true, 00:16:04.665 "data_offset": 0, 00:16:04.665 "data_size": 65536 00:16:04.665 }, 00:16:04.665 { 00:16:04.665 "name": "BaseBdev4", 00:16:04.665 "uuid": "702a66f9-0efb-5830-9bda-b0fa94be140d", 00:16:04.665 "is_configured": true, 00:16:04.665 "data_offset": 0, 00:16:04.665 "data_size": 65536 00:16:04.665 } 00:16:04.665 ] 00:16:04.665 }' 00:16:04.665 15:21:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:04.665 15:21:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:05.234 15:21:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:05.234 15:21:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:16:05.234 15:21:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:05.234 15:21:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:05.234 [2024-11-19 15:21:55.343769] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:05.234 15:21:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:05.234 15:21:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=196608 00:16:05.234 15:21:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:05.234 15:21:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:05.234 15:21:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:05.234 15:21:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:16:05.234 15:21:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:05.234 15:21:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:16:05.234 15:21:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:16:05.234 15:21:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:16:05.234 15:21:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:16:05.234 15:21:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:16:05.234 15:21:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:16:05.234 15:21:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:16:05.234 15:21:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:16:05.234 15:21:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:16:05.234 15:21:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:16:05.234 15:21:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:16:05.234 15:21:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:16:05.234 15:21:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:16:05.234 15:21:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:16:05.494 [2024-11-19 15:21:55.615178] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:16:05.494 /dev/nbd0 00:16:05.494 15:21:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:16:05.494 15:21:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:16:05.494 15:21:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:16:05.494 15:21:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:16:05.494 15:21:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:16:05.494 15:21:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:16:05.494 15:21:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:16:05.494 15:21:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@877 -- # break 00:16:05.494 15:21:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:16:05.494 15:21:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:16:05.494 15:21:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:05.494 1+0 records in 00:16:05.494 1+0 records out 00:16:05.494 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000516885 s, 7.9 MB/s 00:16:05.494 15:21:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:05.494 15:21:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:16:05.494 15:21:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:05.494 15:21:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:16:05.494 15:21:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:16:05.494 15:21:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:16:05.494 15:21:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:16:05.494 15:21:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:16:05.494 15:21:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@630 -- # write_unit_size=384 00:16:05.494 15:21:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@631 -- # echo 192 00:16:05.494 15:21:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=196608 count=512 oflag=direct 00:16:06.064 512+0 records in 00:16:06.064 512+0 records out 00:16:06.064 100663296 bytes (101 MB, 96 MiB) copied, 0.555739 s, 181 MB/s 00:16:06.064 15:21:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:16:06.064 15:21:56 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:16:06.064 15:21:56 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:16:06.064 15:21:56 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:16:06.064 15:21:56 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:16:06.064 15:21:56 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:06.064 15:21:56 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:16:06.324 15:21:56 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:16:06.324 [2024-11-19 15:21:56.470073] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:06.324 15:21:56 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:16:06.324 15:21:56 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:16:06.324 15:21:56 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:06.324 15:21:56 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:06.324 15:21:56 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:16:06.324 15:21:56 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:16:06.324 15:21:56 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:16:06.324 15:21:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:16:06.324 15:21:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:06.324 15:21:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:06.324 [2024-11-19 15:21:56.490135] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:16:06.324 15:21:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:06.324 15:21:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:16:06.324 15:21:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:06.324 15:21:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:06.324 15:21:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:16:06.324 15:21:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:16:06.324 15:21:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:16:06.325 15:21:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:06.325 15:21:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:06.325 15:21:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:06.325 15:21:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:06.325 15:21:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:06.325 15:21:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:06.325 15:21:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:06.325 15:21:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:06.325 15:21:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:06.325 15:21:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:06.325 "name": "raid_bdev1", 00:16:06.325 "uuid": "9539b9e4-8c91-4dcd-8499-03e3c8ef07ea", 00:16:06.325 "strip_size_kb": 64, 00:16:06.325 "state": "online", 00:16:06.325 "raid_level": "raid5f", 00:16:06.325 "superblock": false, 00:16:06.325 "num_base_bdevs": 4, 00:16:06.325 "num_base_bdevs_discovered": 3, 00:16:06.325 "num_base_bdevs_operational": 3, 00:16:06.325 "base_bdevs_list": [ 00:16:06.325 { 00:16:06.325 "name": null, 00:16:06.325 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:06.325 "is_configured": false, 00:16:06.325 "data_offset": 0, 00:16:06.325 "data_size": 65536 00:16:06.325 }, 00:16:06.325 { 00:16:06.325 "name": "BaseBdev2", 00:16:06.325 "uuid": "06197ab8-31a3-55c0-893d-cbd85e6e3e17", 00:16:06.325 "is_configured": true, 00:16:06.325 "data_offset": 0, 00:16:06.325 "data_size": 65536 00:16:06.325 }, 00:16:06.325 { 00:16:06.325 "name": "BaseBdev3", 00:16:06.325 "uuid": "0163c5e4-ac32-5fd6-a6cc-5260e8058920", 00:16:06.325 "is_configured": true, 00:16:06.325 "data_offset": 0, 00:16:06.325 "data_size": 65536 00:16:06.325 }, 00:16:06.325 { 00:16:06.325 "name": "BaseBdev4", 00:16:06.325 "uuid": "702a66f9-0efb-5830-9bda-b0fa94be140d", 00:16:06.325 "is_configured": true, 00:16:06.325 "data_offset": 0, 00:16:06.325 "data_size": 65536 00:16:06.325 } 00:16:06.325 ] 00:16:06.325 }' 00:16:06.325 15:21:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:06.325 15:21:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:06.584 15:21:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:06.584 15:21:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:06.584 15:21:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:06.584 [2024-11-19 15:21:56.917401] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:06.584 [2024-11-19 15:21:56.921758] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000027da0 00:16:06.843 15:21:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:06.843 15:21:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:16:06.843 [2024-11-19 15:21:56.924079] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:07.780 15:21:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:07.780 15:21:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:07.780 15:21:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:07.780 15:21:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:07.780 15:21:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:07.780 15:21:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:07.780 15:21:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:07.780 15:21:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:07.780 15:21:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:07.780 15:21:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:07.780 15:21:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:07.780 "name": "raid_bdev1", 00:16:07.780 "uuid": "9539b9e4-8c91-4dcd-8499-03e3c8ef07ea", 00:16:07.780 "strip_size_kb": 64, 00:16:07.780 "state": "online", 00:16:07.780 "raid_level": "raid5f", 00:16:07.780 "superblock": false, 00:16:07.780 "num_base_bdevs": 4, 00:16:07.780 "num_base_bdevs_discovered": 4, 00:16:07.780 "num_base_bdevs_operational": 4, 00:16:07.780 "process": { 00:16:07.780 "type": "rebuild", 00:16:07.780 "target": "spare", 00:16:07.780 "progress": { 00:16:07.780 "blocks": 19200, 00:16:07.780 "percent": 9 00:16:07.780 } 00:16:07.780 }, 00:16:07.780 "base_bdevs_list": [ 00:16:07.780 { 00:16:07.780 "name": "spare", 00:16:07.780 "uuid": "6dbe7dda-5dfa-5325-a12e-5205861fcefe", 00:16:07.780 "is_configured": true, 00:16:07.780 "data_offset": 0, 00:16:07.780 "data_size": 65536 00:16:07.780 }, 00:16:07.780 { 00:16:07.780 "name": "BaseBdev2", 00:16:07.780 "uuid": "06197ab8-31a3-55c0-893d-cbd85e6e3e17", 00:16:07.780 "is_configured": true, 00:16:07.780 "data_offset": 0, 00:16:07.780 "data_size": 65536 00:16:07.780 }, 00:16:07.780 { 00:16:07.780 "name": "BaseBdev3", 00:16:07.780 "uuid": "0163c5e4-ac32-5fd6-a6cc-5260e8058920", 00:16:07.780 "is_configured": true, 00:16:07.780 "data_offset": 0, 00:16:07.780 "data_size": 65536 00:16:07.780 }, 00:16:07.780 { 00:16:07.780 "name": "BaseBdev4", 00:16:07.780 "uuid": "702a66f9-0efb-5830-9bda-b0fa94be140d", 00:16:07.780 "is_configured": true, 00:16:07.780 "data_offset": 0, 00:16:07.780 "data_size": 65536 00:16:07.780 } 00:16:07.780 ] 00:16:07.780 }' 00:16:07.780 15:21:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:07.780 15:21:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:07.780 15:21:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:07.780 15:21:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:07.780 15:21:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:16:07.780 15:21:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:07.780 15:21:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:07.780 [2024-11-19 15:21:58.064515] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:08.039 [2024-11-19 15:21:58.129690] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:08.039 [2024-11-19 15:21:58.129742] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:08.039 [2024-11-19 15:21:58.129761] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:08.039 [2024-11-19 15:21:58.129768] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:08.039 15:21:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:08.039 15:21:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:16:08.039 15:21:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:08.039 15:21:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:08.039 15:21:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:16:08.039 15:21:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:16:08.039 15:21:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:16:08.039 15:21:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:08.039 15:21:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:08.039 15:21:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:08.039 15:21:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:08.039 15:21:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:08.039 15:21:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:08.039 15:21:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:08.039 15:21:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:08.039 15:21:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:08.039 15:21:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:08.039 "name": "raid_bdev1", 00:16:08.039 "uuid": "9539b9e4-8c91-4dcd-8499-03e3c8ef07ea", 00:16:08.039 "strip_size_kb": 64, 00:16:08.039 "state": "online", 00:16:08.039 "raid_level": "raid5f", 00:16:08.039 "superblock": false, 00:16:08.039 "num_base_bdevs": 4, 00:16:08.039 "num_base_bdevs_discovered": 3, 00:16:08.039 "num_base_bdevs_operational": 3, 00:16:08.039 "base_bdevs_list": [ 00:16:08.039 { 00:16:08.039 "name": null, 00:16:08.039 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:08.039 "is_configured": false, 00:16:08.039 "data_offset": 0, 00:16:08.039 "data_size": 65536 00:16:08.039 }, 00:16:08.039 { 00:16:08.039 "name": "BaseBdev2", 00:16:08.039 "uuid": "06197ab8-31a3-55c0-893d-cbd85e6e3e17", 00:16:08.039 "is_configured": true, 00:16:08.039 "data_offset": 0, 00:16:08.039 "data_size": 65536 00:16:08.039 }, 00:16:08.039 { 00:16:08.039 "name": "BaseBdev3", 00:16:08.039 "uuid": "0163c5e4-ac32-5fd6-a6cc-5260e8058920", 00:16:08.039 "is_configured": true, 00:16:08.039 "data_offset": 0, 00:16:08.039 "data_size": 65536 00:16:08.039 }, 00:16:08.039 { 00:16:08.039 "name": "BaseBdev4", 00:16:08.039 "uuid": "702a66f9-0efb-5830-9bda-b0fa94be140d", 00:16:08.039 "is_configured": true, 00:16:08.039 "data_offset": 0, 00:16:08.039 "data_size": 65536 00:16:08.039 } 00:16:08.039 ] 00:16:08.039 }' 00:16:08.039 15:21:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:08.039 15:21:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:08.298 15:21:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:08.298 15:21:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:08.298 15:21:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:08.298 15:21:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:08.298 15:21:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:08.298 15:21:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:08.298 15:21:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:08.298 15:21:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:08.298 15:21:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:08.298 15:21:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:08.557 15:21:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:08.557 "name": "raid_bdev1", 00:16:08.557 "uuid": "9539b9e4-8c91-4dcd-8499-03e3c8ef07ea", 00:16:08.557 "strip_size_kb": 64, 00:16:08.557 "state": "online", 00:16:08.557 "raid_level": "raid5f", 00:16:08.557 "superblock": false, 00:16:08.557 "num_base_bdevs": 4, 00:16:08.557 "num_base_bdevs_discovered": 3, 00:16:08.557 "num_base_bdevs_operational": 3, 00:16:08.557 "base_bdevs_list": [ 00:16:08.557 { 00:16:08.557 "name": null, 00:16:08.557 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:08.557 "is_configured": false, 00:16:08.557 "data_offset": 0, 00:16:08.557 "data_size": 65536 00:16:08.557 }, 00:16:08.557 { 00:16:08.557 "name": "BaseBdev2", 00:16:08.557 "uuid": "06197ab8-31a3-55c0-893d-cbd85e6e3e17", 00:16:08.557 "is_configured": true, 00:16:08.557 "data_offset": 0, 00:16:08.557 "data_size": 65536 00:16:08.557 }, 00:16:08.557 { 00:16:08.557 "name": "BaseBdev3", 00:16:08.557 "uuid": "0163c5e4-ac32-5fd6-a6cc-5260e8058920", 00:16:08.557 "is_configured": true, 00:16:08.557 "data_offset": 0, 00:16:08.557 "data_size": 65536 00:16:08.557 }, 00:16:08.557 { 00:16:08.557 "name": "BaseBdev4", 00:16:08.557 "uuid": "702a66f9-0efb-5830-9bda-b0fa94be140d", 00:16:08.557 "is_configured": true, 00:16:08.557 "data_offset": 0, 00:16:08.557 "data_size": 65536 00:16:08.557 } 00:16:08.557 ] 00:16:08.557 }' 00:16:08.557 15:21:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:08.557 15:21:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:08.557 15:21:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:08.557 15:21:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:08.557 15:21:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:08.557 15:21:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:08.557 15:21:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:08.557 [2024-11-19 15:21:58.746456] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:08.557 [2024-11-19 15:21:58.750334] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000027e70 00:16:08.557 [2024-11-19 15:21:58.752488] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:08.557 15:21:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:08.557 15:21:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:16:09.496 15:21:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:09.496 15:21:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:09.496 15:21:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:09.496 15:21:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:09.496 15:21:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:09.496 15:21:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:09.496 15:21:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:09.496 15:21:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:09.496 15:21:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:09.496 15:21:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:09.496 15:21:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:09.496 "name": "raid_bdev1", 00:16:09.496 "uuid": "9539b9e4-8c91-4dcd-8499-03e3c8ef07ea", 00:16:09.496 "strip_size_kb": 64, 00:16:09.496 "state": "online", 00:16:09.496 "raid_level": "raid5f", 00:16:09.496 "superblock": false, 00:16:09.496 "num_base_bdevs": 4, 00:16:09.496 "num_base_bdevs_discovered": 4, 00:16:09.496 "num_base_bdevs_operational": 4, 00:16:09.496 "process": { 00:16:09.496 "type": "rebuild", 00:16:09.496 "target": "spare", 00:16:09.496 "progress": { 00:16:09.496 "blocks": 19200, 00:16:09.496 "percent": 9 00:16:09.496 } 00:16:09.496 }, 00:16:09.496 "base_bdevs_list": [ 00:16:09.496 { 00:16:09.496 "name": "spare", 00:16:09.496 "uuid": "6dbe7dda-5dfa-5325-a12e-5205861fcefe", 00:16:09.496 "is_configured": true, 00:16:09.496 "data_offset": 0, 00:16:09.496 "data_size": 65536 00:16:09.496 }, 00:16:09.496 { 00:16:09.496 "name": "BaseBdev2", 00:16:09.496 "uuid": "06197ab8-31a3-55c0-893d-cbd85e6e3e17", 00:16:09.496 "is_configured": true, 00:16:09.496 "data_offset": 0, 00:16:09.496 "data_size": 65536 00:16:09.496 }, 00:16:09.496 { 00:16:09.496 "name": "BaseBdev3", 00:16:09.496 "uuid": "0163c5e4-ac32-5fd6-a6cc-5260e8058920", 00:16:09.496 "is_configured": true, 00:16:09.496 "data_offset": 0, 00:16:09.496 "data_size": 65536 00:16:09.496 }, 00:16:09.496 { 00:16:09.496 "name": "BaseBdev4", 00:16:09.496 "uuid": "702a66f9-0efb-5830-9bda-b0fa94be140d", 00:16:09.496 "is_configured": true, 00:16:09.496 "data_offset": 0, 00:16:09.496 "data_size": 65536 00:16:09.496 } 00:16:09.496 ] 00:16:09.496 }' 00:16:09.496 15:21:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:09.756 15:21:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:09.756 15:21:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:09.756 15:21:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:09.756 15:21:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:16:09.756 15:21:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:16:09.756 15:21:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:16:09.757 15:21:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=517 00:16:09.757 15:21:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:09.757 15:21:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:09.757 15:21:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:09.757 15:21:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:09.757 15:21:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:09.757 15:21:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:09.757 15:21:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:09.757 15:21:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:09.757 15:21:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:09.757 15:21:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:09.757 15:21:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:09.757 15:21:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:09.757 "name": "raid_bdev1", 00:16:09.757 "uuid": "9539b9e4-8c91-4dcd-8499-03e3c8ef07ea", 00:16:09.757 "strip_size_kb": 64, 00:16:09.757 "state": "online", 00:16:09.757 "raid_level": "raid5f", 00:16:09.757 "superblock": false, 00:16:09.757 "num_base_bdevs": 4, 00:16:09.757 "num_base_bdevs_discovered": 4, 00:16:09.757 "num_base_bdevs_operational": 4, 00:16:09.757 "process": { 00:16:09.757 "type": "rebuild", 00:16:09.757 "target": "spare", 00:16:09.757 "progress": { 00:16:09.757 "blocks": 21120, 00:16:09.757 "percent": 10 00:16:09.757 } 00:16:09.757 }, 00:16:09.757 "base_bdevs_list": [ 00:16:09.757 { 00:16:09.757 "name": "spare", 00:16:09.757 "uuid": "6dbe7dda-5dfa-5325-a12e-5205861fcefe", 00:16:09.757 "is_configured": true, 00:16:09.757 "data_offset": 0, 00:16:09.757 "data_size": 65536 00:16:09.757 }, 00:16:09.757 { 00:16:09.757 "name": "BaseBdev2", 00:16:09.757 "uuid": "06197ab8-31a3-55c0-893d-cbd85e6e3e17", 00:16:09.757 "is_configured": true, 00:16:09.757 "data_offset": 0, 00:16:09.757 "data_size": 65536 00:16:09.757 }, 00:16:09.757 { 00:16:09.757 "name": "BaseBdev3", 00:16:09.757 "uuid": "0163c5e4-ac32-5fd6-a6cc-5260e8058920", 00:16:09.757 "is_configured": true, 00:16:09.757 "data_offset": 0, 00:16:09.757 "data_size": 65536 00:16:09.757 }, 00:16:09.757 { 00:16:09.757 "name": "BaseBdev4", 00:16:09.757 "uuid": "702a66f9-0efb-5830-9bda-b0fa94be140d", 00:16:09.757 "is_configured": true, 00:16:09.757 "data_offset": 0, 00:16:09.757 "data_size": 65536 00:16:09.757 } 00:16:09.757 ] 00:16:09.757 }' 00:16:09.757 15:21:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:09.757 15:21:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:09.757 15:21:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:09.757 15:22:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:09.757 15:22:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:10.697 15:22:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:10.697 15:22:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:10.697 15:22:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:10.697 15:22:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:10.697 15:22:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:10.697 15:22:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:10.697 15:22:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:10.697 15:22:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:10.697 15:22:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:10.697 15:22:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:10.957 15:22:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:10.957 15:22:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:10.957 "name": "raid_bdev1", 00:16:10.957 "uuid": "9539b9e4-8c91-4dcd-8499-03e3c8ef07ea", 00:16:10.957 "strip_size_kb": 64, 00:16:10.957 "state": "online", 00:16:10.957 "raid_level": "raid5f", 00:16:10.957 "superblock": false, 00:16:10.957 "num_base_bdevs": 4, 00:16:10.957 "num_base_bdevs_discovered": 4, 00:16:10.957 "num_base_bdevs_operational": 4, 00:16:10.957 "process": { 00:16:10.957 "type": "rebuild", 00:16:10.957 "target": "spare", 00:16:10.957 "progress": { 00:16:10.957 "blocks": 42240, 00:16:10.957 "percent": 21 00:16:10.957 } 00:16:10.957 }, 00:16:10.957 "base_bdevs_list": [ 00:16:10.957 { 00:16:10.957 "name": "spare", 00:16:10.957 "uuid": "6dbe7dda-5dfa-5325-a12e-5205861fcefe", 00:16:10.957 "is_configured": true, 00:16:10.957 "data_offset": 0, 00:16:10.957 "data_size": 65536 00:16:10.957 }, 00:16:10.957 { 00:16:10.957 "name": "BaseBdev2", 00:16:10.957 "uuid": "06197ab8-31a3-55c0-893d-cbd85e6e3e17", 00:16:10.957 "is_configured": true, 00:16:10.957 "data_offset": 0, 00:16:10.957 "data_size": 65536 00:16:10.957 }, 00:16:10.957 { 00:16:10.957 "name": "BaseBdev3", 00:16:10.957 "uuid": "0163c5e4-ac32-5fd6-a6cc-5260e8058920", 00:16:10.957 "is_configured": true, 00:16:10.957 "data_offset": 0, 00:16:10.957 "data_size": 65536 00:16:10.957 }, 00:16:10.957 { 00:16:10.957 "name": "BaseBdev4", 00:16:10.957 "uuid": "702a66f9-0efb-5830-9bda-b0fa94be140d", 00:16:10.957 "is_configured": true, 00:16:10.957 "data_offset": 0, 00:16:10.957 "data_size": 65536 00:16:10.957 } 00:16:10.957 ] 00:16:10.957 }' 00:16:10.957 15:22:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:10.957 15:22:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:10.957 15:22:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:10.957 15:22:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:10.958 15:22:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:11.899 15:22:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:11.899 15:22:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:11.899 15:22:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:11.899 15:22:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:11.899 15:22:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:11.899 15:22:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:11.899 15:22:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:11.899 15:22:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:11.899 15:22:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:11.899 15:22:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:11.899 15:22:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:11.899 15:22:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:11.899 "name": "raid_bdev1", 00:16:11.899 "uuid": "9539b9e4-8c91-4dcd-8499-03e3c8ef07ea", 00:16:11.899 "strip_size_kb": 64, 00:16:11.899 "state": "online", 00:16:11.899 "raid_level": "raid5f", 00:16:11.899 "superblock": false, 00:16:11.899 "num_base_bdevs": 4, 00:16:11.899 "num_base_bdevs_discovered": 4, 00:16:11.899 "num_base_bdevs_operational": 4, 00:16:11.899 "process": { 00:16:11.899 "type": "rebuild", 00:16:11.899 "target": "spare", 00:16:11.899 "progress": { 00:16:11.899 "blocks": 65280, 00:16:11.899 "percent": 33 00:16:11.899 } 00:16:11.899 }, 00:16:11.899 "base_bdevs_list": [ 00:16:11.899 { 00:16:11.899 "name": "spare", 00:16:11.899 "uuid": "6dbe7dda-5dfa-5325-a12e-5205861fcefe", 00:16:11.899 "is_configured": true, 00:16:11.899 "data_offset": 0, 00:16:11.899 "data_size": 65536 00:16:11.899 }, 00:16:11.899 { 00:16:11.899 "name": "BaseBdev2", 00:16:11.899 "uuid": "06197ab8-31a3-55c0-893d-cbd85e6e3e17", 00:16:11.899 "is_configured": true, 00:16:11.899 "data_offset": 0, 00:16:11.899 "data_size": 65536 00:16:11.899 }, 00:16:11.899 { 00:16:11.899 "name": "BaseBdev3", 00:16:11.899 "uuid": "0163c5e4-ac32-5fd6-a6cc-5260e8058920", 00:16:11.899 "is_configured": true, 00:16:11.899 "data_offset": 0, 00:16:11.899 "data_size": 65536 00:16:11.899 }, 00:16:11.899 { 00:16:11.899 "name": "BaseBdev4", 00:16:11.899 "uuid": "702a66f9-0efb-5830-9bda-b0fa94be140d", 00:16:11.899 "is_configured": true, 00:16:11.899 "data_offset": 0, 00:16:11.899 "data_size": 65536 00:16:11.899 } 00:16:11.899 ] 00:16:11.899 }' 00:16:11.899 15:22:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:12.160 15:22:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:12.160 15:22:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:12.160 15:22:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:12.160 15:22:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:13.102 15:22:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:13.102 15:22:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:13.102 15:22:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:13.102 15:22:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:13.102 15:22:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:13.102 15:22:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:13.102 15:22:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:13.102 15:22:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:13.102 15:22:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:13.102 15:22:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:13.102 15:22:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:13.102 15:22:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:13.102 "name": "raid_bdev1", 00:16:13.102 "uuid": "9539b9e4-8c91-4dcd-8499-03e3c8ef07ea", 00:16:13.102 "strip_size_kb": 64, 00:16:13.102 "state": "online", 00:16:13.102 "raid_level": "raid5f", 00:16:13.102 "superblock": false, 00:16:13.102 "num_base_bdevs": 4, 00:16:13.102 "num_base_bdevs_discovered": 4, 00:16:13.102 "num_base_bdevs_operational": 4, 00:16:13.102 "process": { 00:16:13.102 "type": "rebuild", 00:16:13.102 "target": "spare", 00:16:13.102 "progress": { 00:16:13.102 "blocks": 86400, 00:16:13.102 "percent": 43 00:16:13.102 } 00:16:13.102 }, 00:16:13.102 "base_bdevs_list": [ 00:16:13.102 { 00:16:13.102 "name": "spare", 00:16:13.102 "uuid": "6dbe7dda-5dfa-5325-a12e-5205861fcefe", 00:16:13.102 "is_configured": true, 00:16:13.102 "data_offset": 0, 00:16:13.102 "data_size": 65536 00:16:13.102 }, 00:16:13.102 { 00:16:13.102 "name": "BaseBdev2", 00:16:13.102 "uuid": "06197ab8-31a3-55c0-893d-cbd85e6e3e17", 00:16:13.102 "is_configured": true, 00:16:13.102 "data_offset": 0, 00:16:13.102 "data_size": 65536 00:16:13.102 }, 00:16:13.102 { 00:16:13.102 "name": "BaseBdev3", 00:16:13.102 "uuid": "0163c5e4-ac32-5fd6-a6cc-5260e8058920", 00:16:13.102 "is_configured": true, 00:16:13.102 "data_offset": 0, 00:16:13.102 "data_size": 65536 00:16:13.102 }, 00:16:13.102 { 00:16:13.102 "name": "BaseBdev4", 00:16:13.102 "uuid": "702a66f9-0efb-5830-9bda-b0fa94be140d", 00:16:13.102 "is_configured": true, 00:16:13.102 "data_offset": 0, 00:16:13.102 "data_size": 65536 00:16:13.102 } 00:16:13.102 ] 00:16:13.102 }' 00:16:13.102 15:22:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:13.102 15:22:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:13.102 15:22:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:13.362 15:22:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:13.362 15:22:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:14.304 15:22:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:14.304 15:22:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:14.304 15:22:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:14.304 15:22:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:14.304 15:22:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:14.304 15:22:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:14.304 15:22:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:14.304 15:22:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:14.304 15:22:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:14.304 15:22:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:14.304 15:22:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:14.304 15:22:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:14.304 "name": "raid_bdev1", 00:16:14.304 "uuid": "9539b9e4-8c91-4dcd-8499-03e3c8ef07ea", 00:16:14.304 "strip_size_kb": 64, 00:16:14.304 "state": "online", 00:16:14.304 "raid_level": "raid5f", 00:16:14.304 "superblock": false, 00:16:14.304 "num_base_bdevs": 4, 00:16:14.304 "num_base_bdevs_discovered": 4, 00:16:14.304 "num_base_bdevs_operational": 4, 00:16:14.304 "process": { 00:16:14.304 "type": "rebuild", 00:16:14.304 "target": "spare", 00:16:14.304 "progress": { 00:16:14.304 "blocks": 107520, 00:16:14.304 "percent": 54 00:16:14.304 } 00:16:14.304 }, 00:16:14.304 "base_bdevs_list": [ 00:16:14.304 { 00:16:14.304 "name": "spare", 00:16:14.304 "uuid": "6dbe7dda-5dfa-5325-a12e-5205861fcefe", 00:16:14.304 "is_configured": true, 00:16:14.304 "data_offset": 0, 00:16:14.304 "data_size": 65536 00:16:14.304 }, 00:16:14.304 { 00:16:14.304 "name": "BaseBdev2", 00:16:14.304 "uuid": "06197ab8-31a3-55c0-893d-cbd85e6e3e17", 00:16:14.304 "is_configured": true, 00:16:14.304 "data_offset": 0, 00:16:14.304 "data_size": 65536 00:16:14.304 }, 00:16:14.304 { 00:16:14.304 "name": "BaseBdev3", 00:16:14.304 "uuid": "0163c5e4-ac32-5fd6-a6cc-5260e8058920", 00:16:14.304 "is_configured": true, 00:16:14.304 "data_offset": 0, 00:16:14.304 "data_size": 65536 00:16:14.304 }, 00:16:14.304 { 00:16:14.304 "name": "BaseBdev4", 00:16:14.304 "uuid": "702a66f9-0efb-5830-9bda-b0fa94be140d", 00:16:14.304 "is_configured": true, 00:16:14.304 "data_offset": 0, 00:16:14.304 "data_size": 65536 00:16:14.304 } 00:16:14.304 ] 00:16:14.304 }' 00:16:14.304 15:22:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:14.304 15:22:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:14.304 15:22:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:14.304 15:22:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:14.304 15:22:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:15.245 15:22:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:15.245 15:22:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:15.245 15:22:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:15.245 15:22:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:15.245 15:22:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:15.245 15:22:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:15.505 15:22:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:15.505 15:22:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:15.505 15:22:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:15.505 15:22:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:15.505 15:22:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:15.505 15:22:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:15.505 "name": "raid_bdev1", 00:16:15.505 "uuid": "9539b9e4-8c91-4dcd-8499-03e3c8ef07ea", 00:16:15.505 "strip_size_kb": 64, 00:16:15.505 "state": "online", 00:16:15.505 "raid_level": "raid5f", 00:16:15.505 "superblock": false, 00:16:15.505 "num_base_bdevs": 4, 00:16:15.505 "num_base_bdevs_discovered": 4, 00:16:15.505 "num_base_bdevs_operational": 4, 00:16:15.505 "process": { 00:16:15.505 "type": "rebuild", 00:16:15.505 "target": "spare", 00:16:15.505 "progress": { 00:16:15.505 "blocks": 130560, 00:16:15.505 "percent": 66 00:16:15.505 } 00:16:15.505 }, 00:16:15.505 "base_bdevs_list": [ 00:16:15.505 { 00:16:15.505 "name": "spare", 00:16:15.505 "uuid": "6dbe7dda-5dfa-5325-a12e-5205861fcefe", 00:16:15.505 "is_configured": true, 00:16:15.505 "data_offset": 0, 00:16:15.505 "data_size": 65536 00:16:15.505 }, 00:16:15.505 { 00:16:15.505 "name": "BaseBdev2", 00:16:15.505 "uuid": "06197ab8-31a3-55c0-893d-cbd85e6e3e17", 00:16:15.505 "is_configured": true, 00:16:15.505 "data_offset": 0, 00:16:15.505 "data_size": 65536 00:16:15.505 }, 00:16:15.505 { 00:16:15.505 "name": "BaseBdev3", 00:16:15.505 "uuid": "0163c5e4-ac32-5fd6-a6cc-5260e8058920", 00:16:15.505 "is_configured": true, 00:16:15.505 "data_offset": 0, 00:16:15.505 "data_size": 65536 00:16:15.505 }, 00:16:15.505 { 00:16:15.505 "name": "BaseBdev4", 00:16:15.505 "uuid": "702a66f9-0efb-5830-9bda-b0fa94be140d", 00:16:15.505 "is_configured": true, 00:16:15.505 "data_offset": 0, 00:16:15.505 "data_size": 65536 00:16:15.505 } 00:16:15.505 ] 00:16:15.505 }' 00:16:15.505 15:22:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:15.505 15:22:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:15.505 15:22:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:15.505 15:22:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:15.505 15:22:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:16.446 15:22:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:16.446 15:22:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:16.446 15:22:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:16.446 15:22:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:16.446 15:22:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:16.446 15:22:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:16.446 15:22:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:16.446 15:22:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:16.446 15:22:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:16.446 15:22:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:16.446 15:22:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:16.707 15:22:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:16.707 "name": "raid_bdev1", 00:16:16.707 "uuid": "9539b9e4-8c91-4dcd-8499-03e3c8ef07ea", 00:16:16.707 "strip_size_kb": 64, 00:16:16.707 "state": "online", 00:16:16.707 "raid_level": "raid5f", 00:16:16.707 "superblock": false, 00:16:16.707 "num_base_bdevs": 4, 00:16:16.707 "num_base_bdevs_discovered": 4, 00:16:16.707 "num_base_bdevs_operational": 4, 00:16:16.707 "process": { 00:16:16.707 "type": "rebuild", 00:16:16.707 "target": "spare", 00:16:16.707 "progress": { 00:16:16.707 "blocks": 151680, 00:16:16.707 "percent": 77 00:16:16.707 } 00:16:16.707 }, 00:16:16.707 "base_bdevs_list": [ 00:16:16.707 { 00:16:16.707 "name": "spare", 00:16:16.707 "uuid": "6dbe7dda-5dfa-5325-a12e-5205861fcefe", 00:16:16.707 "is_configured": true, 00:16:16.707 "data_offset": 0, 00:16:16.707 "data_size": 65536 00:16:16.707 }, 00:16:16.707 { 00:16:16.707 "name": "BaseBdev2", 00:16:16.707 "uuid": "06197ab8-31a3-55c0-893d-cbd85e6e3e17", 00:16:16.707 "is_configured": true, 00:16:16.707 "data_offset": 0, 00:16:16.707 "data_size": 65536 00:16:16.707 }, 00:16:16.707 { 00:16:16.707 "name": "BaseBdev3", 00:16:16.707 "uuid": "0163c5e4-ac32-5fd6-a6cc-5260e8058920", 00:16:16.707 "is_configured": true, 00:16:16.707 "data_offset": 0, 00:16:16.707 "data_size": 65536 00:16:16.707 }, 00:16:16.707 { 00:16:16.707 "name": "BaseBdev4", 00:16:16.707 "uuid": "702a66f9-0efb-5830-9bda-b0fa94be140d", 00:16:16.707 "is_configured": true, 00:16:16.707 "data_offset": 0, 00:16:16.707 "data_size": 65536 00:16:16.707 } 00:16:16.707 ] 00:16:16.707 }' 00:16:16.707 15:22:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:16.707 15:22:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:16.707 15:22:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:16.707 15:22:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:16.707 15:22:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:17.647 15:22:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:17.647 15:22:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:17.647 15:22:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:17.647 15:22:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:17.647 15:22:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:17.647 15:22:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:17.647 15:22:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:17.647 15:22:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:17.647 15:22:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:17.647 15:22:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:17.648 15:22:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:17.648 15:22:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:17.648 "name": "raid_bdev1", 00:16:17.648 "uuid": "9539b9e4-8c91-4dcd-8499-03e3c8ef07ea", 00:16:17.648 "strip_size_kb": 64, 00:16:17.648 "state": "online", 00:16:17.648 "raid_level": "raid5f", 00:16:17.648 "superblock": false, 00:16:17.648 "num_base_bdevs": 4, 00:16:17.648 "num_base_bdevs_discovered": 4, 00:16:17.648 "num_base_bdevs_operational": 4, 00:16:17.648 "process": { 00:16:17.648 "type": "rebuild", 00:16:17.648 "target": "spare", 00:16:17.648 "progress": { 00:16:17.648 "blocks": 172800, 00:16:17.648 "percent": 87 00:16:17.648 } 00:16:17.648 }, 00:16:17.648 "base_bdevs_list": [ 00:16:17.648 { 00:16:17.648 "name": "spare", 00:16:17.648 "uuid": "6dbe7dda-5dfa-5325-a12e-5205861fcefe", 00:16:17.648 "is_configured": true, 00:16:17.648 "data_offset": 0, 00:16:17.648 "data_size": 65536 00:16:17.648 }, 00:16:17.648 { 00:16:17.648 "name": "BaseBdev2", 00:16:17.648 "uuid": "06197ab8-31a3-55c0-893d-cbd85e6e3e17", 00:16:17.648 "is_configured": true, 00:16:17.648 "data_offset": 0, 00:16:17.648 "data_size": 65536 00:16:17.648 }, 00:16:17.648 { 00:16:17.648 "name": "BaseBdev3", 00:16:17.648 "uuid": "0163c5e4-ac32-5fd6-a6cc-5260e8058920", 00:16:17.648 "is_configured": true, 00:16:17.648 "data_offset": 0, 00:16:17.648 "data_size": 65536 00:16:17.648 }, 00:16:17.648 { 00:16:17.648 "name": "BaseBdev4", 00:16:17.648 "uuid": "702a66f9-0efb-5830-9bda-b0fa94be140d", 00:16:17.648 "is_configured": true, 00:16:17.648 "data_offset": 0, 00:16:17.648 "data_size": 65536 00:16:17.648 } 00:16:17.648 ] 00:16:17.648 }' 00:16:17.648 15:22:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:17.648 15:22:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:17.648 15:22:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:17.908 15:22:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:17.908 15:22:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:18.849 15:22:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:18.849 15:22:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:18.849 15:22:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:18.849 15:22:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:18.849 15:22:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:18.849 15:22:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:18.849 15:22:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:18.849 15:22:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:18.849 15:22:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:18.849 15:22:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:18.849 15:22:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:18.849 15:22:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:18.849 "name": "raid_bdev1", 00:16:18.849 "uuid": "9539b9e4-8c91-4dcd-8499-03e3c8ef07ea", 00:16:18.849 "strip_size_kb": 64, 00:16:18.849 "state": "online", 00:16:18.849 "raid_level": "raid5f", 00:16:18.849 "superblock": false, 00:16:18.849 "num_base_bdevs": 4, 00:16:18.849 "num_base_bdevs_discovered": 4, 00:16:18.849 "num_base_bdevs_operational": 4, 00:16:18.849 "process": { 00:16:18.849 "type": "rebuild", 00:16:18.849 "target": "spare", 00:16:18.849 "progress": { 00:16:18.849 "blocks": 195840, 00:16:18.849 "percent": 99 00:16:18.849 } 00:16:18.849 }, 00:16:18.849 "base_bdevs_list": [ 00:16:18.849 { 00:16:18.849 "name": "spare", 00:16:18.849 "uuid": "6dbe7dda-5dfa-5325-a12e-5205861fcefe", 00:16:18.849 "is_configured": true, 00:16:18.849 "data_offset": 0, 00:16:18.849 "data_size": 65536 00:16:18.849 }, 00:16:18.849 { 00:16:18.849 "name": "BaseBdev2", 00:16:18.849 "uuid": "06197ab8-31a3-55c0-893d-cbd85e6e3e17", 00:16:18.849 "is_configured": true, 00:16:18.849 "data_offset": 0, 00:16:18.849 "data_size": 65536 00:16:18.849 }, 00:16:18.849 { 00:16:18.849 "name": "BaseBdev3", 00:16:18.849 "uuid": "0163c5e4-ac32-5fd6-a6cc-5260e8058920", 00:16:18.849 "is_configured": true, 00:16:18.849 "data_offset": 0, 00:16:18.849 "data_size": 65536 00:16:18.849 }, 00:16:18.849 { 00:16:18.849 "name": "BaseBdev4", 00:16:18.849 "uuid": "702a66f9-0efb-5830-9bda-b0fa94be140d", 00:16:18.849 "is_configured": true, 00:16:18.849 "data_offset": 0, 00:16:18.849 "data_size": 65536 00:16:18.849 } 00:16:18.849 ] 00:16:18.849 }' 00:16:18.849 15:22:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:18.849 [2024-11-19 15:22:09.091711] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:16:18.849 [2024-11-19 15:22:09.091780] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:16:18.849 [2024-11-19 15:22:09.091824] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:18.849 15:22:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:18.849 15:22:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:18.849 15:22:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:18.849 15:22:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:20.229 "name": "raid_bdev1", 00:16:20.229 "uuid": "9539b9e4-8c91-4dcd-8499-03e3c8ef07ea", 00:16:20.229 "strip_size_kb": 64, 00:16:20.229 "state": "online", 00:16:20.229 "raid_level": "raid5f", 00:16:20.229 "superblock": false, 00:16:20.229 "num_base_bdevs": 4, 00:16:20.229 "num_base_bdevs_discovered": 4, 00:16:20.229 "num_base_bdevs_operational": 4, 00:16:20.229 "base_bdevs_list": [ 00:16:20.229 { 00:16:20.229 "name": "spare", 00:16:20.229 "uuid": "6dbe7dda-5dfa-5325-a12e-5205861fcefe", 00:16:20.229 "is_configured": true, 00:16:20.229 "data_offset": 0, 00:16:20.229 "data_size": 65536 00:16:20.229 }, 00:16:20.229 { 00:16:20.229 "name": "BaseBdev2", 00:16:20.229 "uuid": "06197ab8-31a3-55c0-893d-cbd85e6e3e17", 00:16:20.229 "is_configured": true, 00:16:20.229 "data_offset": 0, 00:16:20.229 "data_size": 65536 00:16:20.229 }, 00:16:20.229 { 00:16:20.229 "name": "BaseBdev3", 00:16:20.229 "uuid": "0163c5e4-ac32-5fd6-a6cc-5260e8058920", 00:16:20.229 "is_configured": true, 00:16:20.229 "data_offset": 0, 00:16:20.229 "data_size": 65536 00:16:20.229 }, 00:16:20.229 { 00:16:20.229 "name": "BaseBdev4", 00:16:20.229 "uuid": "702a66f9-0efb-5830-9bda-b0fa94be140d", 00:16:20.229 "is_configured": true, 00:16:20.229 "data_offset": 0, 00:16:20.229 "data_size": 65536 00:16:20.229 } 00:16:20.229 ] 00:16:20.229 }' 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:20.229 "name": "raid_bdev1", 00:16:20.229 "uuid": "9539b9e4-8c91-4dcd-8499-03e3c8ef07ea", 00:16:20.229 "strip_size_kb": 64, 00:16:20.229 "state": "online", 00:16:20.229 "raid_level": "raid5f", 00:16:20.229 "superblock": false, 00:16:20.229 "num_base_bdevs": 4, 00:16:20.229 "num_base_bdevs_discovered": 4, 00:16:20.229 "num_base_bdevs_operational": 4, 00:16:20.229 "base_bdevs_list": [ 00:16:20.229 { 00:16:20.229 "name": "spare", 00:16:20.229 "uuid": "6dbe7dda-5dfa-5325-a12e-5205861fcefe", 00:16:20.229 "is_configured": true, 00:16:20.229 "data_offset": 0, 00:16:20.229 "data_size": 65536 00:16:20.229 }, 00:16:20.229 { 00:16:20.229 "name": "BaseBdev2", 00:16:20.229 "uuid": "06197ab8-31a3-55c0-893d-cbd85e6e3e17", 00:16:20.229 "is_configured": true, 00:16:20.229 "data_offset": 0, 00:16:20.229 "data_size": 65536 00:16:20.229 }, 00:16:20.229 { 00:16:20.229 "name": "BaseBdev3", 00:16:20.229 "uuid": "0163c5e4-ac32-5fd6-a6cc-5260e8058920", 00:16:20.229 "is_configured": true, 00:16:20.229 "data_offset": 0, 00:16:20.229 "data_size": 65536 00:16:20.229 }, 00:16:20.229 { 00:16:20.229 "name": "BaseBdev4", 00:16:20.229 "uuid": "702a66f9-0efb-5830-9bda-b0fa94be140d", 00:16:20.229 "is_configured": true, 00:16:20.229 "data_offset": 0, 00:16:20.229 "data_size": 65536 00:16:20.229 } 00:16:20.229 ] 00:16:20.229 }' 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:20.229 15:22:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:20.230 15:22:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:20.230 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:20.230 "name": "raid_bdev1", 00:16:20.230 "uuid": "9539b9e4-8c91-4dcd-8499-03e3c8ef07ea", 00:16:20.230 "strip_size_kb": 64, 00:16:20.230 "state": "online", 00:16:20.230 "raid_level": "raid5f", 00:16:20.230 "superblock": false, 00:16:20.230 "num_base_bdevs": 4, 00:16:20.230 "num_base_bdevs_discovered": 4, 00:16:20.230 "num_base_bdevs_operational": 4, 00:16:20.230 "base_bdevs_list": [ 00:16:20.230 { 00:16:20.230 "name": "spare", 00:16:20.230 "uuid": "6dbe7dda-5dfa-5325-a12e-5205861fcefe", 00:16:20.230 "is_configured": true, 00:16:20.230 "data_offset": 0, 00:16:20.230 "data_size": 65536 00:16:20.230 }, 00:16:20.230 { 00:16:20.230 "name": "BaseBdev2", 00:16:20.230 "uuid": "06197ab8-31a3-55c0-893d-cbd85e6e3e17", 00:16:20.230 "is_configured": true, 00:16:20.230 "data_offset": 0, 00:16:20.230 "data_size": 65536 00:16:20.230 }, 00:16:20.230 { 00:16:20.230 "name": "BaseBdev3", 00:16:20.230 "uuid": "0163c5e4-ac32-5fd6-a6cc-5260e8058920", 00:16:20.230 "is_configured": true, 00:16:20.230 "data_offset": 0, 00:16:20.230 "data_size": 65536 00:16:20.230 }, 00:16:20.230 { 00:16:20.230 "name": "BaseBdev4", 00:16:20.230 "uuid": "702a66f9-0efb-5830-9bda-b0fa94be140d", 00:16:20.230 "is_configured": true, 00:16:20.230 "data_offset": 0, 00:16:20.230 "data_size": 65536 00:16:20.230 } 00:16:20.230 ] 00:16:20.230 }' 00:16:20.230 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:20.230 15:22:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:20.800 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:20.800 15:22:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:20.800 15:22:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:20.800 [2024-11-19 15:22:10.910180] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:20.800 [2024-11-19 15:22:10.910256] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:20.800 [2024-11-19 15:22:10.910352] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:20.800 [2024-11-19 15:22:10.910479] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:20.800 [2024-11-19 15:22:10.910534] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:16:20.800 15:22:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:20.800 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:20.800 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:16:20.800 15:22:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:20.800 15:22:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:20.800 15:22:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:20.800 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:16:20.800 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:16:20.800 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:16:20.800 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:16:20.800 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:16:20.800 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:16:20.800 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:16:20.800 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:16:20.800 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:16:20.800 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:16:20.800 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:16:20.800 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:16:20.800 15:22:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:16:21.061 /dev/nbd0 00:16:21.061 15:22:11 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:16:21.061 15:22:11 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:16:21.061 15:22:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:16:21.061 15:22:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:16:21.061 15:22:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:16:21.061 15:22:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:16:21.061 15:22:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:16:21.061 15:22:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@877 -- # break 00:16:21.061 15:22:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:16:21.061 15:22:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:16:21.061 15:22:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:21.061 1+0 records in 00:16:21.061 1+0 records out 00:16:21.061 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000527214 s, 7.8 MB/s 00:16:21.061 15:22:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:21.061 15:22:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:16:21.061 15:22:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:21.061 15:22:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:16:21.061 15:22:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:16:21.061 15:22:11 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:16:21.061 15:22:11 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:16:21.061 15:22:11 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:16:21.322 /dev/nbd1 00:16:21.322 15:22:11 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:16:21.322 15:22:11 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:16:21.322 15:22:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:16:21.322 15:22:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:16:21.322 15:22:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:16:21.322 15:22:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:16:21.322 15:22:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:16:21.322 15:22:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@877 -- # break 00:16:21.322 15:22:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:16:21.322 15:22:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:16:21.322 15:22:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:21.322 1+0 records in 00:16:21.322 1+0 records out 00:16:21.322 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000478873 s, 8.6 MB/s 00:16:21.322 15:22:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:21.322 15:22:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:16:21.322 15:22:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:21.322 15:22:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:16:21.322 15:22:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:16:21.322 15:22:11 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:16:21.322 15:22:11 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:16:21.322 15:22:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:16:21.322 15:22:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:16:21.322 15:22:11 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:16:21.322 15:22:11 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:16:21.322 15:22:11 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:16:21.322 15:22:11 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:16:21.322 15:22:11 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:21.322 15:22:11 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:16:21.582 15:22:11 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:16:21.582 15:22:11 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:16:21.582 15:22:11 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:16:21.582 15:22:11 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:21.582 15:22:11 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:21.582 15:22:11 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:16:21.582 15:22:11 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:16:21.582 15:22:11 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:16:21.582 15:22:11 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:21.583 15:22:11 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:16:21.843 15:22:11 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:16:21.843 15:22:11 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:16:21.843 15:22:11 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:16:21.843 15:22:11 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:21.843 15:22:11 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:21.843 15:22:11 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:16:21.843 15:22:11 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:16:21.843 15:22:11 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:16:21.843 15:22:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:16:21.843 15:22:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 95024 00:16:21.843 15:22:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@954 -- # '[' -z 95024 ']' 00:16:21.843 15:22:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@958 -- # kill -0 95024 00:16:21.843 15:22:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@959 -- # uname 00:16:21.843 15:22:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:16:21.843 15:22:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 95024 00:16:21.843 15:22:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:16:21.843 killing process with pid 95024 00:16:21.843 Received shutdown signal, test time was about 60.000000 seconds 00:16:21.843 00:16:21.843 Latency(us) 00:16:21.843 [2024-11-19T15:22:12.182Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:16:21.843 [2024-11-19T15:22:12.182Z] =================================================================================================================== 00:16:21.843 [2024-11-19T15:22:12.182Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:16:21.843 15:22:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:16:21.843 15:22:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 95024' 00:16:21.843 15:22:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@973 -- # kill 95024 00:16:21.843 [2024-11-19 15:22:12.018631] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:21.843 15:22:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@978 -- # wait 95024 00:16:21.843 [2024-11-19 15:22:12.070425] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:22.103 15:22:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:16:22.103 00:16:22.103 real 0m18.454s 00:16:22.103 user 0m22.134s 00:16:22.103 sys 0m2.557s 00:16:22.103 15:22:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:22.103 15:22:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:16:22.103 ************************************ 00:16:22.103 END TEST raid5f_rebuild_test 00:16:22.103 ************************************ 00:16:22.103 15:22:12 bdev_raid -- bdev/bdev_raid.sh@991 -- # run_test raid5f_rebuild_test_sb raid_rebuild_test raid5f 4 true false true 00:16:22.103 15:22:12 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:16:22.103 15:22:12 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:22.103 15:22:12 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:22.103 ************************************ 00:16:22.103 START TEST raid5f_rebuild_test_sb 00:16:22.103 ************************************ 00:16:22.103 15:22:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid5f 4 true false true 00:16:22.103 15:22:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:16:22.103 15:22:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:16:22.103 15:22:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:16:22.103 15:22:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:16:22.103 15:22:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:16:22.103 15:22:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:16:22.103 15:22:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:22.104 15:22:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:16:22.104 15:22:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:16:22.104 15:22:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:22.104 15:22:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:16:22.104 15:22:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:16:22.104 15:22:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:22.104 15:22:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:16:22.104 15:22:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:16:22.104 15:22:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:22.104 15:22:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:16:22.104 15:22:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:16:22.104 15:22:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:22.104 15:22:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:16:22.104 15:22:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:16:22.104 15:22:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:16:22.104 15:22:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:16:22.104 15:22:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:16:22.104 15:22:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:16:22.104 15:22:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:16:22.104 15:22:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:16:22.104 15:22:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:16:22.104 15:22:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:16:22.104 15:22:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:16:22.104 15:22:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:16:22.104 15:22:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:16:22.104 15:22:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=95530 00:16:22.104 15:22:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:16:22.104 15:22:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 95530 00:16:22.104 15:22:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@835 -- # '[' -z 95530 ']' 00:16:22.104 15:22:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:22.104 15:22:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:16:22.104 15:22:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:22.104 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:22.104 15:22:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:16:22.104 15:22:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:22.364 [2024-11-19 15:22:12.442156] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:16:22.364 [2024-11-19 15:22:12.442333] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.ealI/O size of 3145728 is greater than zero copy threshold (65536). 00:16:22.364 Zero copy mechanism will not be used. 00:16:22.364 :6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid95530 ] 00:16:22.364 [2024-11-19 15:22:12.597565] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:22.364 [2024-11-19 15:22:12.622303] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:22.364 [2024-11-19 15:22:12.665376] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:22.364 [2024-11-19 15:22:12.665497] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:22.940 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:16:22.940 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@868 -- # return 0 00:16:22.940 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:16:22.940 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:16:22.940 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:22.940 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:23.201 BaseBdev1_malloc 00:16:23.201 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:23.201 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:16:23.201 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:23.201 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:23.201 [2024-11-19 15:22:13.285002] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:16:23.201 [2024-11-19 15:22:13.285049] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:23.201 [2024-11-19 15:22:13.285078] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:16:23.201 [2024-11-19 15:22:13.285091] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:23.201 [2024-11-19 15:22:13.287222] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:23.201 [2024-11-19 15:22:13.287260] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:16:23.201 BaseBdev1 00:16:23.201 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:23.201 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:16:23.201 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:16:23.201 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:23.201 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:23.201 BaseBdev2_malloc 00:16:23.201 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:23.201 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:16:23.201 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:23.201 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:23.201 [2024-11-19 15:22:13.313508] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:16:23.201 [2024-11-19 15:22:13.313553] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:23.201 [2024-11-19 15:22:13.313572] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:16:23.201 [2024-11-19 15:22:13.313580] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:23.201 [2024-11-19 15:22:13.315632] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:23.201 [2024-11-19 15:22:13.315671] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:16:23.201 BaseBdev2 00:16:23.201 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:23.201 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:16:23.201 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:16:23.201 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:23.201 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:23.201 BaseBdev3_malloc 00:16:23.201 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:23.201 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:16:23.201 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:23.201 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:23.201 [2024-11-19 15:22:13.342017] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:16:23.201 [2024-11-19 15:22:13.342061] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:23.201 [2024-11-19 15:22:13.342082] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:16:23.201 [2024-11-19 15:22:13.342090] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:23.201 [2024-11-19 15:22:13.344075] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:23.201 [2024-11-19 15:22:13.344110] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:16:23.201 BaseBdev3 00:16:23.201 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:23.201 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:16:23.201 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:16:23.201 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:23.201 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:23.201 BaseBdev4_malloc 00:16:23.201 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:23.201 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:16:23.201 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:23.201 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:23.201 [2024-11-19 15:22:13.388168] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:16:23.202 [2024-11-19 15:22:13.388358] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:23.202 [2024-11-19 15:22:13.388418] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:16:23.202 [2024-11-19 15:22:13.388441] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:23.202 [2024-11-19 15:22:13.392244] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:23.202 [2024-11-19 15:22:13.392296] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:16:23.202 BaseBdev4 00:16:23.202 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:23.202 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:16:23.202 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:23.202 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:23.202 spare_malloc 00:16:23.202 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:23.202 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:16:23.202 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:23.202 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:23.202 spare_delay 00:16:23.202 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:23.202 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:23.202 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:23.202 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:23.202 [2024-11-19 15:22:13.430235] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:23.202 [2024-11-19 15:22:13.430317] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:23.202 [2024-11-19 15:22:13.430339] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:16:23.202 [2024-11-19 15:22:13.430348] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:23.202 [2024-11-19 15:22:13.432378] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:23.202 [2024-11-19 15:22:13.432415] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:23.202 spare 00:16:23.202 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:23.202 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:16:23.202 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:23.202 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:23.202 [2024-11-19 15:22:13.442289] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:23.202 [2024-11-19 15:22:13.444033] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:23.202 [2024-11-19 15:22:13.444094] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:16:23.202 [2024-11-19 15:22:13.444138] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:16:23.202 [2024-11-19 15:22:13.444306] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:16:23.202 [2024-11-19 15:22:13.444317] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:16:23.202 [2024-11-19 15:22:13.444538] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:16:23.202 [2024-11-19 15:22:13.444964] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:16:23.202 [2024-11-19 15:22:13.444996] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:16:23.202 [2024-11-19 15:22:13.445125] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:23.202 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:23.202 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:16:23.202 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:23.202 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:23.202 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:16:23.202 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:16:23.202 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:16:23.202 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:23.202 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:23.202 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:23.202 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:23.202 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:23.202 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:23.202 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:23.202 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:23.202 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:23.202 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:23.202 "name": "raid_bdev1", 00:16:23.202 "uuid": "49213557-9482-4923-a3f8-3e1a78fbbc90", 00:16:23.202 "strip_size_kb": 64, 00:16:23.202 "state": "online", 00:16:23.202 "raid_level": "raid5f", 00:16:23.202 "superblock": true, 00:16:23.202 "num_base_bdevs": 4, 00:16:23.202 "num_base_bdevs_discovered": 4, 00:16:23.202 "num_base_bdevs_operational": 4, 00:16:23.202 "base_bdevs_list": [ 00:16:23.202 { 00:16:23.202 "name": "BaseBdev1", 00:16:23.202 "uuid": "07506305-465d-5826-9d7f-ac0a53ff2b44", 00:16:23.202 "is_configured": true, 00:16:23.202 "data_offset": 2048, 00:16:23.202 "data_size": 63488 00:16:23.202 }, 00:16:23.202 { 00:16:23.202 "name": "BaseBdev2", 00:16:23.202 "uuid": "369db297-95ae-56bd-88b3-690234fd2e40", 00:16:23.202 "is_configured": true, 00:16:23.202 "data_offset": 2048, 00:16:23.202 "data_size": 63488 00:16:23.202 }, 00:16:23.202 { 00:16:23.202 "name": "BaseBdev3", 00:16:23.202 "uuid": "d8f88c70-5105-55ff-b37b-778fcf5056bb", 00:16:23.202 "is_configured": true, 00:16:23.202 "data_offset": 2048, 00:16:23.202 "data_size": 63488 00:16:23.202 }, 00:16:23.202 { 00:16:23.202 "name": "BaseBdev4", 00:16:23.202 "uuid": "c1bdaaee-4f20-5259-9a48-adf4beab217a", 00:16:23.202 "is_configured": true, 00:16:23.202 "data_offset": 2048, 00:16:23.202 "data_size": 63488 00:16:23.202 } 00:16:23.202 ] 00:16:23.202 }' 00:16:23.202 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:23.202 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:23.772 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:23.772 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:23.772 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:23.772 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:16:23.772 [2024-11-19 15:22:13.902235] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:23.772 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:23.772 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=190464 00:16:23.772 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:23.772 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:16:23.772 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:23.772 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:23.772 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:23.772 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:16:23.772 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:16:23.772 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:16:23.772 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:16:23.772 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:16:23.772 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:16:23.772 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:16:23.772 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:16:23.772 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:16:23.772 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:16:23.772 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:16:23.773 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:16:23.773 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:16:23.773 15:22:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:16:24.033 [2024-11-19 15:22:14.157673] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:16:24.033 /dev/nbd0 00:16:24.033 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:16:24.033 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:16:24.033 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:16:24.033 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:16:24.033 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:16:24.033 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:16:24.033 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:16:24.033 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:16:24.033 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:16:24.033 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:16:24.033 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:24.033 1+0 records in 00:16:24.033 1+0 records out 00:16:24.033 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000306704 s, 13.4 MB/s 00:16:24.033 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:24.033 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:16:24.033 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:24.033 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:16:24.033 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:16:24.033 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:16:24.033 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:16:24.033 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:16:24.033 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@630 -- # write_unit_size=384 00:16:24.033 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@631 -- # echo 192 00:16:24.033 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=196608 count=496 oflag=direct 00:16:24.293 496+0 records in 00:16:24.293 496+0 records out 00:16:24.293 97517568 bytes (98 MB, 93 MiB) copied, 0.393751 s, 248 MB/s 00:16:24.293 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:16:24.293 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:16:24.293 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:16:24.293 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:16:24.293 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:16:24.293 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:24.293 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:16:24.553 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:16:24.553 [2024-11-19 15:22:14.851259] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:24.553 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:16:24.553 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:16:24.553 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:24.553 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:24.553 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:16:24.553 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:16:24.553 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:16:24.553 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:16:24.553 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:24.553 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:24.553 [2024-11-19 15:22:14.868871] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:16:24.553 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:24.553 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:16:24.553 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:24.553 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:24.553 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:16:24.553 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:16:24.553 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:16:24.553 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:24.553 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:24.553 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:24.553 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:24.553 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:24.553 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:24.553 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:24.553 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:24.813 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:24.813 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:24.813 "name": "raid_bdev1", 00:16:24.813 "uuid": "49213557-9482-4923-a3f8-3e1a78fbbc90", 00:16:24.813 "strip_size_kb": 64, 00:16:24.813 "state": "online", 00:16:24.813 "raid_level": "raid5f", 00:16:24.813 "superblock": true, 00:16:24.813 "num_base_bdevs": 4, 00:16:24.813 "num_base_bdevs_discovered": 3, 00:16:24.813 "num_base_bdevs_operational": 3, 00:16:24.813 "base_bdevs_list": [ 00:16:24.813 { 00:16:24.813 "name": null, 00:16:24.813 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:24.813 "is_configured": false, 00:16:24.813 "data_offset": 0, 00:16:24.813 "data_size": 63488 00:16:24.813 }, 00:16:24.813 { 00:16:24.813 "name": "BaseBdev2", 00:16:24.813 "uuid": "369db297-95ae-56bd-88b3-690234fd2e40", 00:16:24.813 "is_configured": true, 00:16:24.813 "data_offset": 2048, 00:16:24.813 "data_size": 63488 00:16:24.813 }, 00:16:24.813 { 00:16:24.813 "name": "BaseBdev3", 00:16:24.813 "uuid": "d8f88c70-5105-55ff-b37b-778fcf5056bb", 00:16:24.813 "is_configured": true, 00:16:24.813 "data_offset": 2048, 00:16:24.813 "data_size": 63488 00:16:24.813 }, 00:16:24.813 { 00:16:24.813 "name": "BaseBdev4", 00:16:24.813 "uuid": "c1bdaaee-4f20-5259-9a48-adf4beab217a", 00:16:24.813 "is_configured": true, 00:16:24.813 "data_offset": 2048, 00:16:24.813 "data_size": 63488 00:16:24.813 } 00:16:24.813 ] 00:16:24.813 }' 00:16:24.813 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:24.813 15:22:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:25.073 15:22:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:25.073 15:22:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:25.073 15:22:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:25.073 [2024-11-19 15:22:15.336070] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:25.073 [2024-11-19 15:22:15.340236] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000270a0 00:16:25.073 15:22:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:25.073 15:22:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:16:25.073 [2024-11-19 15:22:15.342518] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:26.013 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:26.013 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:26.013 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:26.013 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:26.013 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:26.273 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:26.273 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:26.273 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:26.273 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:26.273 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:26.273 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:26.273 "name": "raid_bdev1", 00:16:26.273 "uuid": "49213557-9482-4923-a3f8-3e1a78fbbc90", 00:16:26.273 "strip_size_kb": 64, 00:16:26.273 "state": "online", 00:16:26.273 "raid_level": "raid5f", 00:16:26.273 "superblock": true, 00:16:26.273 "num_base_bdevs": 4, 00:16:26.273 "num_base_bdevs_discovered": 4, 00:16:26.273 "num_base_bdevs_operational": 4, 00:16:26.273 "process": { 00:16:26.273 "type": "rebuild", 00:16:26.273 "target": "spare", 00:16:26.273 "progress": { 00:16:26.273 "blocks": 19200, 00:16:26.273 "percent": 10 00:16:26.273 } 00:16:26.273 }, 00:16:26.273 "base_bdevs_list": [ 00:16:26.273 { 00:16:26.273 "name": "spare", 00:16:26.273 "uuid": "e6ab5de9-b8b4-50cf-abfb-aa54d1ea5673", 00:16:26.273 "is_configured": true, 00:16:26.273 "data_offset": 2048, 00:16:26.273 "data_size": 63488 00:16:26.273 }, 00:16:26.273 { 00:16:26.273 "name": "BaseBdev2", 00:16:26.273 "uuid": "369db297-95ae-56bd-88b3-690234fd2e40", 00:16:26.273 "is_configured": true, 00:16:26.273 "data_offset": 2048, 00:16:26.273 "data_size": 63488 00:16:26.273 }, 00:16:26.273 { 00:16:26.273 "name": "BaseBdev3", 00:16:26.273 "uuid": "d8f88c70-5105-55ff-b37b-778fcf5056bb", 00:16:26.273 "is_configured": true, 00:16:26.273 "data_offset": 2048, 00:16:26.273 "data_size": 63488 00:16:26.273 }, 00:16:26.273 { 00:16:26.273 "name": "BaseBdev4", 00:16:26.273 "uuid": "c1bdaaee-4f20-5259-9a48-adf4beab217a", 00:16:26.273 "is_configured": true, 00:16:26.273 "data_offset": 2048, 00:16:26.273 "data_size": 63488 00:16:26.273 } 00:16:26.273 ] 00:16:26.273 }' 00:16:26.273 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:26.273 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:26.273 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:26.273 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:26.273 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:16:26.273 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:26.273 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:26.273 [2024-11-19 15:22:16.507185] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:26.273 [2024-11-19 15:22:16.547772] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:26.273 [2024-11-19 15:22:16.547888] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:26.273 [2024-11-19 15:22:16.547933] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:26.273 [2024-11-19 15:22:16.547979] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:26.273 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:26.273 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:16:26.273 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:26.273 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:26.273 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:16:26.273 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:16:26.273 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:16:26.273 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:26.273 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:26.273 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:26.273 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:26.273 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:26.273 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:26.273 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:26.273 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:26.273 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:26.533 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:26.533 "name": "raid_bdev1", 00:16:26.533 "uuid": "49213557-9482-4923-a3f8-3e1a78fbbc90", 00:16:26.533 "strip_size_kb": 64, 00:16:26.533 "state": "online", 00:16:26.533 "raid_level": "raid5f", 00:16:26.533 "superblock": true, 00:16:26.533 "num_base_bdevs": 4, 00:16:26.533 "num_base_bdevs_discovered": 3, 00:16:26.533 "num_base_bdevs_operational": 3, 00:16:26.533 "base_bdevs_list": [ 00:16:26.533 { 00:16:26.533 "name": null, 00:16:26.533 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:26.533 "is_configured": false, 00:16:26.533 "data_offset": 0, 00:16:26.533 "data_size": 63488 00:16:26.533 }, 00:16:26.533 { 00:16:26.533 "name": "BaseBdev2", 00:16:26.533 "uuid": "369db297-95ae-56bd-88b3-690234fd2e40", 00:16:26.533 "is_configured": true, 00:16:26.533 "data_offset": 2048, 00:16:26.533 "data_size": 63488 00:16:26.533 }, 00:16:26.533 { 00:16:26.533 "name": "BaseBdev3", 00:16:26.533 "uuid": "d8f88c70-5105-55ff-b37b-778fcf5056bb", 00:16:26.533 "is_configured": true, 00:16:26.533 "data_offset": 2048, 00:16:26.533 "data_size": 63488 00:16:26.533 }, 00:16:26.533 { 00:16:26.533 "name": "BaseBdev4", 00:16:26.533 "uuid": "c1bdaaee-4f20-5259-9a48-adf4beab217a", 00:16:26.533 "is_configured": true, 00:16:26.533 "data_offset": 2048, 00:16:26.533 "data_size": 63488 00:16:26.533 } 00:16:26.533 ] 00:16:26.533 }' 00:16:26.533 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:26.533 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:26.793 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:26.793 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:26.793 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:26.793 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:26.793 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:26.793 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:26.793 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:26.793 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:26.793 15:22:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:26.793 15:22:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:26.793 15:22:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:26.793 "name": "raid_bdev1", 00:16:26.793 "uuid": "49213557-9482-4923-a3f8-3e1a78fbbc90", 00:16:26.793 "strip_size_kb": 64, 00:16:26.793 "state": "online", 00:16:26.793 "raid_level": "raid5f", 00:16:26.793 "superblock": true, 00:16:26.793 "num_base_bdevs": 4, 00:16:26.793 "num_base_bdevs_discovered": 3, 00:16:26.793 "num_base_bdevs_operational": 3, 00:16:26.793 "base_bdevs_list": [ 00:16:26.793 { 00:16:26.793 "name": null, 00:16:26.793 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:26.793 "is_configured": false, 00:16:26.794 "data_offset": 0, 00:16:26.794 "data_size": 63488 00:16:26.794 }, 00:16:26.794 { 00:16:26.794 "name": "BaseBdev2", 00:16:26.794 "uuid": "369db297-95ae-56bd-88b3-690234fd2e40", 00:16:26.794 "is_configured": true, 00:16:26.794 "data_offset": 2048, 00:16:26.794 "data_size": 63488 00:16:26.794 }, 00:16:26.794 { 00:16:26.794 "name": "BaseBdev3", 00:16:26.794 "uuid": "d8f88c70-5105-55ff-b37b-778fcf5056bb", 00:16:26.794 "is_configured": true, 00:16:26.794 "data_offset": 2048, 00:16:26.794 "data_size": 63488 00:16:26.794 }, 00:16:26.794 { 00:16:26.794 "name": "BaseBdev4", 00:16:26.794 "uuid": "c1bdaaee-4f20-5259-9a48-adf4beab217a", 00:16:26.794 "is_configured": true, 00:16:26.794 "data_offset": 2048, 00:16:26.794 "data_size": 63488 00:16:26.794 } 00:16:26.794 ] 00:16:26.794 }' 00:16:26.794 15:22:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:26.794 15:22:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:26.794 15:22:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:26.794 15:22:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:26.794 15:22:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:26.794 15:22:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:26.794 15:22:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:26.794 [2024-11-19 15:22:17.128653] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:27.054 [2024-11-19 15:22:17.132288] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000027170 00:16:27.054 [2024-11-19 15:22:17.134454] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:27.054 15:22:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:27.054 15:22:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:16:27.994 15:22:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:27.994 15:22:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:27.994 15:22:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:27.994 15:22:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:27.994 15:22:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:27.994 15:22:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:27.994 15:22:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:27.994 15:22:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:27.994 15:22:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:27.994 15:22:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:27.994 15:22:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:27.995 "name": "raid_bdev1", 00:16:27.995 "uuid": "49213557-9482-4923-a3f8-3e1a78fbbc90", 00:16:27.995 "strip_size_kb": 64, 00:16:27.995 "state": "online", 00:16:27.995 "raid_level": "raid5f", 00:16:27.995 "superblock": true, 00:16:27.995 "num_base_bdevs": 4, 00:16:27.995 "num_base_bdevs_discovered": 4, 00:16:27.995 "num_base_bdevs_operational": 4, 00:16:27.995 "process": { 00:16:27.995 "type": "rebuild", 00:16:27.995 "target": "spare", 00:16:27.995 "progress": { 00:16:27.995 "blocks": 19200, 00:16:27.995 "percent": 10 00:16:27.995 } 00:16:27.995 }, 00:16:27.995 "base_bdevs_list": [ 00:16:27.995 { 00:16:27.995 "name": "spare", 00:16:27.995 "uuid": "e6ab5de9-b8b4-50cf-abfb-aa54d1ea5673", 00:16:27.995 "is_configured": true, 00:16:27.995 "data_offset": 2048, 00:16:27.995 "data_size": 63488 00:16:27.995 }, 00:16:27.995 { 00:16:27.995 "name": "BaseBdev2", 00:16:27.995 "uuid": "369db297-95ae-56bd-88b3-690234fd2e40", 00:16:27.995 "is_configured": true, 00:16:27.995 "data_offset": 2048, 00:16:27.995 "data_size": 63488 00:16:27.995 }, 00:16:27.995 { 00:16:27.995 "name": "BaseBdev3", 00:16:27.995 "uuid": "d8f88c70-5105-55ff-b37b-778fcf5056bb", 00:16:27.995 "is_configured": true, 00:16:27.995 "data_offset": 2048, 00:16:27.995 "data_size": 63488 00:16:27.995 }, 00:16:27.995 { 00:16:27.995 "name": "BaseBdev4", 00:16:27.995 "uuid": "c1bdaaee-4f20-5259-9a48-adf4beab217a", 00:16:27.995 "is_configured": true, 00:16:27.995 "data_offset": 2048, 00:16:27.995 "data_size": 63488 00:16:27.995 } 00:16:27.995 ] 00:16:27.995 }' 00:16:27.995 15:22:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:27.995 15:22:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:27.995 15:22:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:27.995 15:22:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:27.995 15:22:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:16:27.995 15:22:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:16:27.995 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:16:27.995 15:22:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:16:27.995 15:22:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:16:27.995 15:22:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=536 00:16:27.995 15:22:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:27.995 15:22:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:27.995 15:22:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:27.995 15:22:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:27.995 15:22:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:27.995 15:22:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:27.995 15:22:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:27.995 15:22:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:27.995 15:22:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:27.995 15:22:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:27.995 15:22:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:28.255 15:22:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:28.255 "name": "raid_bdev1", 00:16:28.255 "uuid": "49213557-9482-4923-a3f8-3e1a78fbbc90", 00:16:28.255 "strip_size_kb": 64, 00:16:28.255 "state": "online", 00:16:28.255 "raid_level": "raid5f", 00:16:28.255 "superblock": true, 00:16:28.255 "num_base_bdevs": 4, 00:16:28.255 "num_base_bdevs_discovered": 4, 00:16:28.255 "num_base_bdevs_operational": 4, 00:16:28.255 "process": { 00:16:28.255 "type": "rebuild", 00:16:28.255 "target": "spare", 00:16:28.255 "progress": { 00:16:28.255 "blocks": 21120, 00:16:28.255 "percent": 11 00:16:28.255 } 00:16:28.255 }, 00:16:28.255 "base_bdevs_list": [ 00:16:28.255 { 00:16:28.255 "name": "spare", 00:16:28.255 "uuid": "e6ab5de9-b8b4-50cf-abfb-aa54d1ea5673", 00:16:28.255 "is_configured": true, 00:16:28.255 "data_offset": 2048, 00:16:28.255 "data_size": 63488 00:16:28.255 }, 00:16:28.255 { 00:16:28.255 "name": "BaseBdev2", 00:16:28.255 "uuid": "369db297-95ae-56bd-88b3-690234fd2e40", 00:16:28.255 "is_configured": true, 00:16:28.255 "data_offset": 2048, 00:16:28.255 "data_size": 63488 00:16:28.255 }, 00:16:28.255 { 00:16:28.255 "name": "BaseBdev3", 00:16:28.255 "uuid": "d8f88c70-5105-55ff-b37b-778fcf5056bb", 00:16:28.255 "is_configured": true, 00:16:28.255 "data_offset": 2048, 00:16:28.255 "data_size": 63488 00:16:28.255 }, 00:16:28.255 { 00:16:28.255 "name": "BaseBdev4", 00:16:28.255 "uuid": "c1bdaaee-4f20-5259-9a48-adf4beab217a", 00:16:28.255 "is_configured": true, 00:16:28.255 "data_offset": 2048, 00:16:28.255 "data_size": 63488 00:16:28.255 } 00:16:28.255 ] 00:16:28.255 }' 00:16:28.255 15:22:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:28.255 15:22:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:28.255 15:22:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:28.255 15:22:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:28.255 15:22:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:29.193 15:22:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:29.193 15:22:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:29.193 15:22:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:29.193 15:22:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:29.193 15:22:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:29.193 15:22:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:29.193 15:22:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:29.193 15:22:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:29.193 15:22:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:29.193 15:22:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:29.193 15:22:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:29.193 15:22:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:29.193 "name": "raid_bdev1", 00:16:29.193 "uuid": "49213557-9482-4923-a3f8-3e1a78fbbc90", 00:16:29.193 "strip_size_kb": 64, 00:16:29.193 "state": "online", 00:16:29.193 "raid_level": "raid5f", 00:16:29.193 "superblock": true, 00:16:29.193 "num_base_bdevs": 4, 00:16:29.193 "num_base_bdevs_discovered": 4, 00:16:29.193 "num_base_bdevs_operational": 4, 00:16:29.193 "process": { 00:16:29.193 "type": "rebuild", 00:16:29.193 "target": "spare", 00:16:29.193 "progress": { 00:16:29.193 "blocks": 44160, 00:16:29.193 "percent": 23 00:16:29.193 } 00:16:29.193 }, 00:16:29.193 "base_bdevs_list": [ 00:16:29.193 { 00:16:29.193 "name": "spare", 00:16:29.193 "uuid": "e6ab5de9-b8b4-50cf-abfb-aa54d1ea5673", 00:16:29.193 "is_configured": true, 00:16:29.193 "data_offset": 2048, 00:16:29.193 "data_size": 63488 00:16:29.193 }, 00:16:29.193 { 00:16:29.193 "name": "BaseBdev2", 00:16:29.193 "uuid": "369db297-95ae-56bd-88b3-690234fd2e40", 00:16:29.193 "is_configured": true, 00:16:29.193 "data_offset": 2048, 00:16:29.193 "data_size": 63488 00:16:29.193 }, 00:16:29.193 { 00:16:29.193 "name": "BaseBdev3", 00:16:29.194 "uuid": "d8f88c70-5105-55ff-b37b-778fcf5056bb", 00:16:29.194 "is_configured": true, 00:16:29.194 "data_offset": 2048, 00:16:29.194 "data_size": 63488 00:16:29.194 }, 00:16:29.194 { 00:16:29.194 "name": "BaseBdev4", 00:16:29.194 "uuid": "c1bdaaee-4f20-5259-9a48-adf4beab217a", 00:16:29.194 "is_configured": true, 00:16:29.194 "data_offset": 2048, 00:16:29.194 "data_size": 63488 00:16:29.194 } 00:16:29.194 ] 00:16:29.194 }' 00:16:29.194 15:22:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:29.453 15:22:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:29.453 15:22:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:29.453 15:22:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:29.453 15:22:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:30.393 15:22:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:30.393 15:22:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:30.393 15:22:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:30.393 15:22:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:30.393 15:22:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:30.393 15:22:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:30.393 15:22:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:30.393 15:22:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:30.393 15:22:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:30.393 15:22:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:30.393 15:22:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:30.393 15:22:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:30.393 "name": "raid_bdev1", 00:16:30.393 "uuid": "49213557-9482-4923-a3f8-3e1a78fbbc90", 00:16:30.393 "strip_size_kb": 64, 00:16:30.393 "state": "online", 00:16:30.393 "raid_level": "raid5f", 00:16:30.393 "superblock": true, 00:16:30.393 "num_base_bdevs": 4, 00:16:30.393 "num_base_bdevs_discovered": 4, 00:16:30.393 "num_base_bdevs_operational": 4, 00:16:30.393 "process": { 00:16:30.393 "type": "rebuild", 00:16:30.393 "target": "spare", 00:16:30.394 "progress": { 00:16:30.394 "blocks": 65280, 00:16:30.394 "percent": 34 00:16:30.394 } 00:16:30.394 }, 00:16:30.394 "base_bdevs_list": [ 00:16:30.394 { 00:16:30.394 "name": "spare", 00:16:30.394 "uuid": "e6ab5de9-b8b4-50cf-abfb-aa54d1ea5673", 00:16:30.394 "is_configured": true, 00:16:30.394 "data_offset": 2048, 00:16:30.394 "data_size": 63488 00:16:30.394 }, 00:16:30.394 { 00:16:30.394 "name": "BaseBdev2", 00:16:30.394 "uuid": "369db297-95ae-56bd-88b3-690234fd2e40", 00:16:30.394 "is_configured": true, 00:16:30.394 "data_offset": 2048, 00:16:30.394 "data_size": 63488 00:16:30.394 }, 00:16:30.394 { 00:16:30.394 "name": "BaseBdev3", 00:16:30.394 "uuid": "d8f88c70-5105-55ff-b37b-778fcf5056bb", 00:16:30.394 "is_configured": true, 00:16:30.394 "data_offset": 2048, 00:16:30.394 "data_size": 63488 00:16:30.394 }, 00:16:30.394 { 00:16:30.394 "name": "BaseBdev4", 00:16:30.394 "uuid": "c1bdaaee-4f20-5259-9a48-adf4beab217a", 00:16:30.394 "is_configured": true, 00:16:30.394 "data_offset": 2048, 00:16:30.394 "data_size": 63488 00:16:30.394 } 00:16:30.394 ] 00:16:30.394 }' 00:16:30.394 15:22:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:30.394 15:22:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:30.394 15:22:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:30.394 15:22:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:30.394 15:22:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:31.778 15:22:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:31.778 15:22:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:31.778 15:22:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:31.778 15:22:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:31.778 15:22:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:31.778 15:22:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:31.778 15:22:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:31.778 15:22:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:31.778 15:22:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:31.778 15:22:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:31.778 15:22:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:31.778 15:22:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:31.778 "name": "raid_bdev1", 00:16:31.778 "uuid": "49213557-9482-4923-a3f8-3e1a78fbbc90", 00:16:31.778 "strip_size_kb": 64, 00:16:31.778 "state": "online", 00:16:31.778 "raid_level": "raid5f", 00:16:31.778 "superblock": true, 00:16:31.778 "num_base_bdevs": 4, 00:16:31.778 "num_base_bdevs_discovered": 4, 00:16:31.778 "num_base_bdevs_operational": 4, 00:16:31.778 "process": { 00:16:31.778 "type": "rebuild", 00:16:31.778 "target": "spare", 00:16:31.778 "progress": { 00:16:31.778 "blocks": 86400, 00:16:31.778 "percent": 45 00:16:31.778 } 00:16:31.778 }, 00:16:31.778 "base_bdevs_list": [ 00:16:31.778 { 00:16:31.778 "name": "spare", 00:16:31.778 "uuid": "e6ab5de9-b8b4-50cf-abfb-aa54d1ea5673", 00:16:31.778 "is_configured": true, 00:16:31.778 "data_offset": 2048, 00:16:31.778 "data_size": 63488 00:16:31.778 }, 00:16:31.778 { 00:16:31.778 "name": "BaseBdev2", 00:16:31.778 "uuid": "369db297-95ae-56bd-88b3-690234fd2e40", 00:16:31.778 "is_configured": true, 00:16:31.778 "data_offset": 2048, 00:16:31.778 "data_size": 63488 00:16:31.778 }, 00:16:31.778 { 00:16:31.778 "name": "BaseBdev3", 00:16:31.778 "uuid": "d8f88c70-5105-55ff-b37b-778fcf5056bb", 00:16:31.778 "is_configured": true, 00:16:31.778 "data_offset": 2048, 00:16:31.778 "data_size": 63488 00:16:31.778 }, 00:16:31.778 { 00:16:31.778 "name": "BaseBdev4", 00:16:31.778 "uuid": "c1bdaaee-4f20-5259-9a48-adf4beab217a", 00:16:31.778 "is_configured": true, 00:16:31.778 "data_offset": 2048, 00:16:31.778 "data_size": 63488 00:16:31.778 } 00:16:31.778 ] 00:16:31.778 }' 00:16:31.778 15:22:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:31.778 15:22:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:31.778 15:22:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:31.778 15:22:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:31.778 15:22:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:32.718 15:22:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:32.718 15:22:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:32.718 15:22:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:32.718 15:22:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:32.718 15:22:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:32.718 15:22:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:32.718 15:22:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:32.718 15:22:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:32.718 15:22:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:32.718 15:22:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:32.718 15:22:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:32.719 15:22:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:32.719 "name": "raid_bdev1", 00:16:32.719 "uuid": "49213557-9482-4923-a3f8-3e1a78fbbc90", 00:16:32.719 "strip_size_kb": 64, 00:16:32.719 "state": "online", 00:16:32.719 "raid_level": "raid5f", 00:16:32.719 "superblock": true, 00:16:32.719 "num_base_bdevs": 4, 00:16:32.719 "num_base_bdevs_discovered": 4, 00:16:32.719 "num_base_bdevs_operational": 4, 00:16:32.719 "process": { 00:16:32.719 "type": "rebuild", 00:16:32.719 "target": "spare", 00:16:32.719 "progress": { 00:16:32.719 "blocks": 109440, 00:16:32.719 "percent": 57 00:16:32.719 } 00:16:32.719 }, 00:16:32.719 "base_bdevs_list": [ 00:16:32.719 { 00:16:32.719 "name": "spare", 00:16:32.719 "uuid": "e6ab5de9-b8b4-50cf-abfb-aa54d1ea5673", 00:16:32.719 "is_configured": true, 00:16:32.719 "data_offset": 2048, 00:16:32.719 "data_size": 63488 00:16:32.719 }, 00:16:32.719 { 00:16:32.719 "name": "BaseBdev2", 00:16:32.719 "uuid": "369db297-95ae-56bd-88b3-690234fd2e40", 00:16:32.719 "is_configured": true, 00:16:32.719 "data_offset": 2048, 00:16:32.719 "data_size": 63488 00:16:32.719 }, 00:16:32.719 { 00:16:32.719 "name": "BaseBdev3", 00:16:32.719 "uuid": "d8f88c70-5105-55ff-b37b-778fcf5056bb", 00:16:32.719 "is_configured": true, 00:16:32.719 "data_offset": 2048, 00:16:32.719 "data_size": 63488 00:16:32.719 }, 00:16:32.719 { 00:16:32.719 "name": "BaseBdev4", 00:16:32.719 "uuid": "c1bdaaee-4f20-5259-9a48-adf4beab217a", 00:16:32.719 "is_configured": true, 00:16:32.719 "data_offset": 2048, 00:16:32.719 "data_size": 63488 00:16:32.719 } 00:16:32.719 ] 00:16:32.719 }' 00:16:32.719 15:22:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:32.719 15:22:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:32.719 15:22:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:32.719 15:22:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:32.719 15:22:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:34.101 15:22:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:34.101 15:22:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:34.101 15:22:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:34.101 15:22:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:34.101 15:22:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:34.101 15:22:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:34.101 15:22:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:34.101 15:22:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:34.101 15:22:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:34.101 15:22:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:34.101 15:22:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:34.101 15:22:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:34.101 "name": "raid_bdev1", 00:16:34.101 "uuid": "49213557-9482-4923-a3f8-3e1a78fbbc90", 00:16:34.101 "strip_size_kb": 64, 00:16:34.101 "state": "online", 00:16:34.101 "raid_level": "raid5f", 00:16:34.101 "superblock": true, 00:16:34.101 "num_base_bdevs": 4, 00:16:34.101 "num_base_bdevs_discovered": 4, 00:16:34.101 "num_base_bdevs_operational": 4, 00:16:34.101 "process": { 00:16:34.101 "type": "rebuild", 00:16:34.101 "target": "spare", 00:16:34.101 "progress": { 00:16:34.101 "blocks": 130560, 00:16:34.101 "percent": 68 00:16:34.101 } 00:16:34.101 }, 00:16:34.101 "base_bdevs_list": [ 00:16:34.101 { 00:16:34.101 "name": "spare", 00:16:34.101 "uuid": "e6ab5de9-b8b4-50cf-abfb-aa54d1ea5673", 00:16:34.101 "is_configured": true, 00:16:34.101 "data_offset": 2048, 00:16:34.101 "data_size": 63488 00:16:34.101 }, 00:16:34.101 { 00:16:34.101 "name": "BaseBdev2", 00:16:34.101 "uuid": "369db297-95ae-56bd-88b3-690234fd2e40", 00:16:34.101 "is_configured": true, 00:16:34.101 "data_offset": 2048, 00:16:34.101 "data_size": 63488 00:16:34.101 }, 00:16:34.101 { 00:16:34.101 "name": "BaseBdev3", 00:16:34.101 "uuid": "d8f88c70-5105-55ff-b37b-778fcf5056bb", 00:16:34.101 "is_configured": true, 00:16:34.101 "data_offset": 2048, 00:16:34.101 "data_size": 63488 00:16:34.101 }, 00:16:34.101 { 00:16:34.101 "name": "BaseBdev4", 00:16:34.101 "uuid": "c1bdaaee-4f20-5259-9a48-adf4beab217a", 00:16:34.101 "is_configured": true, 00:16:34.101 "data_offset": 2048, 00:16:34.101 "data_size": 63488 00:16:34.101 } 00:16:34.101 ] 00:16:34.101 }' 00:16:34.101 15:22:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:34.101 15:22:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:34.101 15:22:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:34.101 15:22:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:34.101 15:22:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:35.043 15:22:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:35.043 15:22:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:35.043 15:22:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:35.043 15:22:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:35.043 15:22:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:35.043 15:22:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:35.043 15:22:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:35.043 15:22:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:35.043 15:22:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:35.043 15:22:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:35.043 15:22:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:35.043 15:22:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:35.043 "name": "raid_bdev1", 00:16:35.043 "uuid": "49213557-9482-4923-a3f8-3e1a78fbbc90", 00:16:35.043 "strip_size_kb": 64, 00:16:35.043 "state": "online", 00:16:35.043 "raid_level": "raid5f", 00:16:35.043 "superblock": true, 00:16:35.043 "num_base_bdevs": 4, 00:16:35.043 "num_base_bdevs_discovered": 4, 00:16:35.043 "num_base_bdevs_operational": 4, 00:16:35.043 "process": { 00:16:35.043 "type": "rebuild", 00:16:35.043 "target": "spare", 00:16:35.043 "progress": { 00:16:35.043 "blocks": 153600, 00:16:35.043 "percent": 80 00:16:35.043 } 00:16:35.043 }, 00:16:35.043 "base_bdevs_list": [ 00:16:35.043 { 00:16:35.043 "name": "spare", 00:16:35.043 "uuid": "e6ab5de9-b8b4-50cf-abfb-aa54d1ea5673", 00:16:35.043 "is_configured": true, 00:16:35.043 "data_offset": 2048, 00:16:35.043 "data_size": 63488 00:16:35.044 }, 00:16:35.044 { 00:16:35.044 "name": "BaseBdev2", 00:16:35.044 "uuid": "369db297-95ae-56bd-88b3-690234fd2e40", 00:16:35.044 "is_configured": true, 00:16:35.044 "data_offset": 2048, 00:16:35.044 "data_size": 63488 00:16:35.044 }, 00:16:35.044 { 00:16:35.044 "name": "BaseBdev3", 00:16:35.044 "uuid": "d8f88c70-5105-55ff-b37b-778fcf5056bb", 00:16:35.044 "is_configured": true, 00:16:35.044 "data_offset": 2048, 00:16:35.044 "data_size": 63488 00:16:35.044 }, 00:16:35.044 { 00:16:35.044 "name": "BaseBdev4", 00:16:35.044 "uuid": "c1bdaaee-4f20-5259-9a48-adf4beab217a", 00:16:35.044 "is_configured": true, 00:16:35.044 "data_offset": 2048, 00:16:35.044 "data_size": 63488 00:16:35.044 } 00:16:35.044 ] 00:16:35.044 }' 00:16:35.044 15:22:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:35.044 15:22:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:35.044 15:22:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:35.044 15:22:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:35.044 15:22:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:35.984 15:22:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:35.984 15:22:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:35.984 15:22:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:35.984 15:22:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:35.984 15:22:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:35.984 15:22:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:35.984 15:22:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:35.984 15:22:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:35.984 15:22:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:35.984 15:22:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:36.244 15:22:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:36.244 15:22:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:36.244 "name": "raid_bdev1", 00:16:36.244 "uuid": "49213557-9482-4923-a3f8-3e1a78fbbc90", 00:16:36.244 "strip_size_kb": 64, 00:16:36.244 "state": "online", 00:16:36.244 "raid_level": "raid5f", 00:16:36.244 "superblock": true, 00:16:36.244 "num_base_bdevs": 4, 00:16:36.244 "num_base_bdevs_discovered": 4, 00:16:36.244 "num_base_bdevs_operational": 4, 00:16:36.244 "process": { 00:16:36.244 "type": "rebuild", 00:16:36.244 "target": "spare", 00:16:36.244 "progress": { 00:16:36.244 "blocks": 174720, 00:16:36.244 "percent": 91 00:16:36.244 } 00:16:36.244 }, 00:16:36.244 "base_bdevs_list": [ 00:16:36.244 { 00:16:36.244 "name": "spare", 00:16:36.244 "uuid": "e6ab5de9-b8b4-50cf-abfb-aa54d1ea5673", 00:16:36.244 "is_configured": true, 00:16:36.244 "data_offset": 2048, 00:16:36.244 "data_size": 63488 00:16:36.244 }, 00:16:36.244 { 00:16:36.244 "name": "BaseBdev2", 00:16:36.244 "uuid": "369db297-95ae-56bd-88b3-690234fd2e40", 00:16:36.244 "is_configured": true, 00:16:36.244 "data_offset": 2048, 00:16:36.244 "data_size": 63488 00:16:36.244 }, 00:16:36.244 { 00:16:36.244 "name": "BaseBdev3", 00:16:36.244 "uuid": "d8f88c70-5105-55ff-b37b-778fcf5056bb", 00:16:36.244 "is_configured": true, 00:16:36.244 "data_offset": 2048, 00:16:36.244 "data_size": 63488 00:16:36.244 }, 00:16:36.244 { 00:16:36.244 "name": "BaseBdev4", 00:16:36.244 "uuid": "c1bdaaee-4f20-5259-9a48-adf4beab217a", 00:16:36.244 "is_configured": true, 00:16:36.244 "data_offset": 2048, 00:16:36.244 "data_size": 63488 00:16:36.244 } 00:16:36.244 ] 00:16:36.244 }' 00:16:36.244 15:22:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:36.244 15:22:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:36.244 15:22:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:36.244 15:22:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:36.244 15:22:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:37.184 [2024-11-19 15:22:27.172391] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:16:37.184 [2024-11-19 15:22:27.172460] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:16:37.184 [2024-11-19 15:22:27.172568] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:37.184 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:37.184 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:37.184 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:37.184 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:37.184 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:37.184 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:37.184 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:37.184 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:37.184 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:37.184 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:37.184 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:37.184 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:37.184 "name": "raid_bdev1", 00:16:37.184 "uuid": "49213557-9482-4923-a3f8-3e1a78fbbc90", 00:16:37.184 "strip_size_kb": 64, 00:16:37.184 "state": "online", 00:16:37.184 "raid_level": "raid5f", 00:16:37.184 "superblock": true, 00:16:37.184 "num_base_bdevs": 4, 00:16:37.184 "num_base_bdevs_discovered": 4, 00:16:37.184 "num_base_bdevs_operational": 4, 00:16:37.184 "base_bdevs_list": [ 00:16:37.184 { 00:16:37.184 "name": "spare", 00:16:37.184 "uuid": "e6ab5de9-b8b4-50cf-abfb-aa54d1ea5673", 00:16:37.184 "is_configured": true, 00:16:37.184 "data_offset": 2048, 00:16:37.184 "data_size": 63488 00:16:37.184 }, 00:16:37.184 { 00:16:37.184 "name": "BaseBdev2", 00:16:37.184 "uuid": "369db297-95ae-56bd-88b3-690234fd2e40", 00:16:37.184 "is_configured": true, 00:16:37.185 "data_offset": 2048, 00:16:37.185 "data_size": 63488 00:16:37.185 }, 00:16:37.185 { 00:16:37.185 "name": "BaseBdev3", 00:16:37.185 "uuid": "d8f88c70-5105-55ff-b37b-778fcf5056bb", 00:16:37.185 "is_configured": true, 00:16:37.185 "data_offset": 2048, 00:16:37.185 "data_size": 63488 00:16:37.185 }, 00:16:37.185 { 00:16:37.185 "name": "BaseBdev4", 00:16:37.185 "uuid": "c1bdaaee-4f20-5259-9a48-adf4beab217a", 00:16:37.185 "is_configured": true, 00:16:37.185 "data_offset": 2048, 00:16:37.185 "data_size": 63488 00:16:37.185 } 00:16:37.185 ] 00:16:37.185 }' 00:16:37.185 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:37.445 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:16:37.445 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:37.445 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:16:37.445 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:16:37.445 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:37.445 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:37.445 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:37.445 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:37.445 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:37.445 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:37.445 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:37.445 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:37.445 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:37.445 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:37.445 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:37.445 "name": "raid_bdev1", 00:16:37.445 "uuid": "49213557-9482-4923-a3f8-3e1a78fbbc90", 00:16:37.445 "strip_size_kb": 64, 00:16:37.445 "state": "online", 00:16:37.445 "raid_level": "raid5f", 00:16:37.445 "superblock": true, 00:16:37.445 "num_base_bdevs": 4, 00:16:37.445 "num_base_bdevs_discovered": 4, 00:16:37.445 "num_base_bdevs_operational": 4, 00:16:37.445 "base_bdevs_list": [ 00:16:37.445 { 00:16:37.445 "name": "spare", 00:16:37.445 "uuid": "e6ab5de9-b8b4-50cf-abfb-aa54d1ea5673", 00:16:37.445 "is_configured": true, 00:16:37.445 "data_offset": 2048, 00:16:37.445 "data_size": 63488 00:16:37.445 }, 00:16:37.445 { 00:16:37.445 "name": "BaseBdev2", 00:16:37.445 "uuid": "369db297-95ae-56bd-88b3-690234fd2e40", 00:16:37.445 "is_configured": true, 00:16:37.445 "data_offset": 2048, 00:16:37.445 "data_size": 63488 00:16:37.445 }, 00:16:37.445 { 00:16:37.445 "name": "BaseBdev3", 00:16:37.445 "uuid": "d8f88c70-5105-55ff-b37b-778fcf5056bb", 00:16:37.445 "is_configured": true, 00:16:37.445 "data_offset": 2048, 00:16:37.445 "data_size": 63488 00:16:37.445 }, 00:16:37.445 { 00:16:37.445 "name": "BaseBdev4", 00:16:37.445 "uuid": "c1bdaaee-4f20-5259-9a48-adf4beab217a", 00:16:37.445 "is_configured": true, 00:16:37.445 "data_offset": 2048, 00:16:37.445 "data_size": 63488 00:16:37.445 } 00:16:37.445 ] 00:16:37.445 }' 00:16:37.445 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:37.445 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:37.445 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:37.445 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:37.445 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:16:37.445 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:37.445 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:37.445 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:16:37.445 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:16:37.445 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:16:37.445 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:37.445 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:37.445 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:37.445 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:37.445 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:37.445 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:37.445 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:37.445 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:37.705 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:37.705 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:37.705 "name": "raid_bdev1", 00:16:37.705 "uuid": "49213557-9482-4923-a3f8-3e1a78fbbc90", 00:16:37.705 "strip_size_kb": 64, 00:16:37.705 "state": "online", 00:16:37.705 "raid_level": "raid5f", 00:16:37.705 "superblock": true, 00:16:37.705 "num_base_bdevs": 4, 00:16:37.705 "num_base_bdevs_discovered": 4, 00:16:37.705 "num_base_bdevs_operational": 4, 00:16:37.705 "base_bdevs_list": [ 00:16:37.705 { 00:16:37.705 "name": "spare", 00:16:37.705 "uuid": "e6ab5de9-b8b4-50cf-abfb-aa54d1ea5673", 00:16:37.705 "is_configured": true, 00:16:37.705 "data_offset": 2048, 00:16:37.705 "data_size": 63488 00:16:37.705 }, 00:16:37.705 { 00:16:37.705 "name": "BaseBdev2", 00:16:37.705 "uuid": "369db297-95ae-56bd-88b3-690234fd2e40", 00:16:37.705 "is_configured": true, 00:16:37.705 "data_offset": 2048, 00:16:37.705 "data_size": 63488 00:16:37.705 }, 00:16:37.705 { 00:16:37.705 "name": "BaseBdev3", 00:16:37.705 "uuid": "d8f88c70-5105-55ff-b37b-778fcf5056bb", 00:16:37.705 "is_configured": true, 00:16:37.705 "data_offset": 2048, 00:16:37.705 "data_size": 63488 00:16:37.705 }, 00:16:37.705 { 00:16:37.705 "name": "BaseBdev4", 00:16:37.705 "uuid": "c1bdaaee-4f20-5259-9a48-adf4beab217a", 00:16:37.705 "is_configured": true, 00:16:37.705 "data_offset": 2048, 00:16:37.705 "data_size": 63488 00:16:37.705 } 00:16:37.705 ] 00:16:37.705 }' 00:16:37.705 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:37.705 15:22:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:37.965 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:37.965 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:37.965 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:37.965 [2024-11-19 15:22:28.248160] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:37.965 [2024-11-19 15:22:28.248232] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:37.965 [2024-11-19 15:22:28.248316] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:37.965 [2024-11-19 15:22:28.248422] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:37.965 [2024-11-19 15:22:28.248447] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:16:37.965 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:37.965 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:37.965 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:16:37.965 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:37.965 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:37.965 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:37.965 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:16:38.226 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:16:38.226 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:16:38.226 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:16:38.226 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:16:38.226 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:16:38.226 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:16:38.226 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:16:38.226 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:16:38.226 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:16:38.226 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:16:38.226 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:16:38.226 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:16:38.226 /dev/nbd0 00:16:38.226 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:16:38.226 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:16:38.226 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:16:38.226 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:16:38.226 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:16:38.226 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:16:38.226 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:16:38.226 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:16:38.226 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:16:38.226 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:16:38.226 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:38.226 1+0 records in 00:16:38.226 1+0 records out 00:16:38.226 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00055426 s, 7.4 MB/s 00:16:38.226 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:38.226 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:16:38.486 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:38.486 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:16:38.486 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:16:38.486 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:16:38.486 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:16:38.486 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:16:38.486 /dev/nbd1 00:16:38.486 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:16:38.486 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:16:38.486 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:16:38.486 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:16:38.486 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:16:38.486 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:16:38.486 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:16:38.486 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:16:38.486 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:16:38.486 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:16:38.486 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:38.486 1+0 records in 00:16:38.486 1+0 records out 00:16:38.486 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000478899 s, 8.6 MB/s 00:16:38.486 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:38.486 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:16:38.486 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:38.487 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:16:38.487 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:16:38.487 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:16:38.487 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:16:38.487 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:16:38.747 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:16:38.747 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:16:38.747 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:16:38.747 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:16:38.747 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:16:38.747 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:38.747 15:22:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:16:38.747 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:16:38.747 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:16:38.747 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:16:38.747 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:38.747 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:38.747 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:16:39.008 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:16:39.008 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:16:39.008 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:39.008 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:16:39.008 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:16:39.008 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:16:39.008 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:16:39.008 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:39.008 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:39.008 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:16:39.008 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:16:39.008 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:16:39.008 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:16:39.008 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:16:39.008 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:39.008 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:39.008 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:39.008 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:39.008 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:39.008 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:39.008 [2024-11-19 15:22:29.334906] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:39.008 [2024-11-19 15:22:29.334979] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:39.008 [2024-11-19 15:22:29.335019] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:16:39.008 [2024-11-19 15:22:29.335034] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:39.008 [2024-11-19 15:22:29.337283] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:39.008 [2024-11-19 15:22:29.337325] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:39.008 [2024-11-19 15:22:29.337396] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:16:39.008 [2024-11-19 15:22:29.337441] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:39.008 [2024-11-19 15:22:29.337556] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:39.008 [2024-11-19 15:22:29.337652] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:16:39.008 [2024-11-19 15:22:29.337714] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:16:39.008 spare 00:16:39.008 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:39.008 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:16:39.008 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:39.008 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:39.268 [2024-11-19 15:22:29.437602] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:16:39.268 [2024-11-19 15:22:29.437632] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:16:39.268 [2024-11-19 15:22:29.437882] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000045820 00:16:39.268 [2024-11-19 15:22:29.438330] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:16:39.268 [2024-11-19 15:22:29.438344] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:16:39.268 [2024-11-19 15:22:29.438458] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:39.268 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:39.268 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:16:39.268 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:39.268 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:39.268 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:16:39.268 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:16:39.268 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:16:39.268 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:39.268 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:39.268 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:39.268 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:39.268 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:39.268 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:39.268 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:39.268 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:39.268 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:39.268 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:39.268 "name": "raid_bdev1", 00:16:39.268 "uuid": "49213557-9482-4923-a3f8-3e1a78fbbc90", 00:16:39.268 "strip_size_kb": 64, 00:16:39.268 "state": "online", 00:16:39.268 "raid_level": "raid5f", 00:16:39.268 "superblock": true, 00:16:39.268 "num_base_bdevs": 4, 00:16:39.268 "num_base_bdevs_discovered": 4, 00:16:39.268 "num_base_bdevs_operational": 4, 00:16:39.268 "base_bdevs_list": [ 00:16:39.268 { 00:16:39.268 "name": "spare", 00:16:39.268 "uuid": "e6ab5de9-b8b4-50cf-abfb-aa54d1ea5673", 00:16:39.268 "is_configured": true, 00:16:39.268 "data_offset": 2048, 00:16:39.268 "data_size": 63488 00:16:39.268 }, 00:16:39.268 { 00:16:39.268 "name": "BaseBdev2", 00:16:39.268 "uuid": "369db297-95ae-56bd-88b3-690234fd2e40", 00:16:39.268 "is_configured": true, 00:16:39.268 "data_offset": 2048, 00:16:39.268 "data_size": 63488 00:16:39.268 }, 00:16:39.268 { 00:16:39.268 "name": "BaseBdev3", 00:16:39.268 "uuid": "d8f88c70-5105-55ff-b37b-778fcf5056bb", 00:16:39.269 "is_configured": true, 00:16:39.269 "data_offset": 2048, 00:16:39.269 "data_size": 63488 00:16:39.269 }, 00:16:39.269 { 00:16:39.269 "name": "BaseBdev4", 00:16:39.269 "uuid": "c1bdaaee-4f20-5259-9a48-adf4beab217a", 00:16:39.269 "is_configured": true, 00:16:39.269 "data_offset": 2048, 00:16:39.269 "data_size": 63488 00:16:39.269 } 00:16:39.269 ] 00:16:39.269 }' 00:16:39.269 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:39.269 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:39.839 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:39.839 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:39.839 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:39.839 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:39.839 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:39.839 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:39.839 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:39.839 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:39.839 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:39.839 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:39.839 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:39.839 "name": "raid_bdev1", 00:16:39.839 "uuid": "49213557-9482-4923-a3f8-3e1a78fbbc90", 00:16:39.839 "strip_size_kb": 64, 00:16:39.839 "state": "online", 00:16:39.839 "raid_level": "raid5f", 00:16:39.839 "superblock": true, 00:16:39.839 "num_base_bdevs": 4, 00:16:39.839 "num_base_bdevs_discovered": 4, 00:16:39.839 "num_base_bdevs_operational": 4, 00:16:39.839 "base_bdevs_list": [ 00:16:39.839 { 00:16:39.839 "name": "spare", 00:16:39.839 "uuid": "e6ab5de9-b8b4-50cf-abfb-aa54d1ea5673", 00:16:39.839 "is_configured": true, 00:16:39.839 "data_offset": 2048, 00:16:39.839 "data_size": 63488 00:16:39.839 }, 00:16:39.839 { 00:16:39.839 "name": "BaseBdev2", 00:16:39.839 "uuid": "369db297-95ae-56bd-88b3-690234fd2e40", 00:16:39.839 "is_configured": true, 00:16:39.839 "data_offset": 2048, 00:16:39.839 "data_size": 63488 00:16:39.839 }, 00:16:39.839 { 00:16:39.839 "name": "BaseBdev3", 00:16:39.839 "uuid": "d8f88c70-5105-55ff-b37b-778fcf5056bb", 00:16:39.839 "is_configured": true, 00:16:39.839 "data_offset": 2048, 00:16:39.839 "data_size": 63488 00:16:39.839 }, 00:16:39.839 { 00:16:39.839 "name": "BaseBdev4", 00:16:39.839 "uuid": "c1bdaaee-4f20-5259-9a48-adf4beab217a", 00:16:39.839 "is_configured": true, 00:16:39.839 "data_offset": 2048, 00:16:39.839 "data_size": 63488 00:16:39.839 } 00:16:39.839 ] 00:16:39.839 }' 00:16:39.839 15:22:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:39.839 15:22:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:39.839 15:22:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:39.839 15:22:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:39.839 15:22:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:39.839 15:22:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:39.839 15:22:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:39.839 15:22:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:16:39.839 15:22:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:39.839 15:22:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:16:39.839 15:22:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:16:39.839 15:22:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:39.839 15:22:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:39.839 [2024-11-19 15:22:30.111021] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:39.839 15:22:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:39.839 15:22:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:16:39.839 15:22:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:39.839 15:22:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:39.839 15:22:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:16:39.839 15:22:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:16:39.839 15:22:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:16:39.839 15:22:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:39.839 15:22:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:39.839 15:22:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:39.839 15:22:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:39.839 15:22:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:39.839 15:22:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:39.839 15:22:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:39.839 15:22:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:39.839 15:22:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:39.839 15:22:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:39.839 "name": "raid_bdev1", 00:16:39.839 "uuid": "49213557-9482-4923-a3f8-3e1a78fbbc90", 00:16:39.839 "strip_size_kb": 64, 00:16:39.839 "state": "online", 00:16:39.839 "raid_level": "raid5f", 00:16:39.839 "superblock": true, 00:16:39.839 "num_base_bdevs": 4, 00:16:39.839 "num_base_bdevs_discovered": 3, 00:16:39.839 "num_base_bdevs_operational": 3, 00:16:39.839 "base_bdevs_list": [ 00:16:39.839 { 00:16:39.839 "name": null, 00:16:39.839 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:39.839 "is_configured": false, 00:16:39.839 "data_offset": 0, 00:16:39.839 "data_size": 63488 00:16:39.839 }, 00:16:39.839 { 00:16:39.839 "name": "BaseBdev2", 00:16:39.839 "uuid": "369db297-95ae-56bd-88b3-690234fd2e40", 00:16:39.839 "is_configured": true, 00:16:39.839 "data_offset": 2048, 00:16:39.839 "data_size": 63488 00:16:39.839 }, 00:16:39.839 { 00:16:39.839 "name": "BaseBdev3", 00:16:39.839 "uuid": "d8f88c70-5105-55ff-b37b-778fcf5056bb", 00:16:39.839 "is_configured": true, 00:16:39.839 "data_offset": 2048, 00:16:39.839 "data_size": 63488 00:16:39.839 }, 00:16:39.839 { 00:16:39.839 "name": "BaseBdev4", 00:16:39.839 "uuid": "c1bdaaee-4f20-5259-9a48-adf4beab217a", 00:16:39.839 "is_configured": true, 00:16:39.839 "data_offset": 2048, 00:16:39.839 "data_size": 63488 00:16:39.839 } 00:16:39.839 ] 00:16:39.839 }' 00:16:39.839 15:22:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:39.839 15:22:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:40.410 15:22:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:40.410 15:22:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:40.410 15:22:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:40.410 [2024-11-19 15:22:30.546258] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:40.410 [2024-11-19 15:22:30.546452] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:16:40.410 [2024-11-19 15:22:30.546494] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:16:40.410 [2024-11-19 15:22:30.546545] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:40.410 [2024-11-19 15:22:30.550649] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000458f0 00:16:40.410 15:22:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:40.410 15:22:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:16:40.410 [2024-11-19 15:22:30.552827] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:41.350 15:22:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:41.350 15:22:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:41.350 15:22:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:41.350 15:22:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:41.350 15:22:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:41.350 15:22:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:41.350 15:22:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:41.350 15:22:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:41.350 15:22:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:41.350 15:22:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:41.350 15:22:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:41.350 "name": "raid_bdev1", 00:16:41.350 "uuid": "49213557-9482-4923-a3f8-3e1a78fbbc90", 00:16:41.350 "strip_size_kb": 64, 00:16:41.350 "state": "online", 00:16:41.350 "raid_level": "raid5f", 00:16:41.350 "superblock": true, 00:16:41.350 "num_base_bdevs": 4, 00:16:41.350 "num_base_bdevs_discovered": 4, 00:16:41.350 "num_base_bdevs_operational": 4, 00:16:41.350 "process": { 00:16:41.350 "type": "rebuild", 00:16:41.350 "target": "spare", 00:16:41.350 "progress": { 00:16:41.350 "blocks": 19200, 00:16:41.350 "percent": 10 00:16:41.350 } 00:16:41.350 }, 00:16:41.350 "base_bdevs_list": [ 00:16:41.350 { 00:16:41.350 "name": "spare", 00:16:41.350 "uuid": "e6ab5de9-b8b4-50cf-abfb-aa54d1ea5673", 00:16:41.350 "is_configured": true, 00:16:41.350 "data_offset": 2048, 00:16:41.350 "data_size": 63488 00:16:41.350 }, 00:16:41.350 { 00:16:41.350 "name": "BaseBdev2", 00:16:41.350 "uuid": "369db297-95ae-56bd-88b3-690234fd2e40", 00:16:41.350 "is_configured": true, 00:16:41.350 "data_offset": 2048, 00:16:41.350 "data_size": 63488 00:16:41.350 }, 00:16:41.350 { 00:16:41.350 "name": "BaseBdev3", 00:16:41.350 "uuid": "d8f88c70-5105-55ff-b37b-778fcf5056bb", 00:16:41.350 "is_configured": true, 00:16:41.350 "data_offset": 2048, 00:16:41.350 "data_size": 63488 00:16:41.350 }, 00:16:41.350 { 00:16:41.350 "name": "BaseBdev4", 00:16:41.350 "uuid": "c1bdaaee-4f20-5259-9a48-adf4beab217a", 00:16:41.350 "is_configured": true, 00:16:41.350 "data_offset": 2048, 00:16:41.350 "data_size": 63488 00:16:41.350 } 00:16:41.350 ] 00:16:41.350 }' 00:16:41.350 15:22:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:41.350 15:22:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:41.350 15:22:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:41.610 15:22:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:41.610 15:22:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:16:41.610 15:22:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:41.610 15:22:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:41.610 [2024-11-19 15:22:31.717327] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:41.610 [2024-11-19 15:22:31.757875] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:41.610 [2024-11-19 15:22:31.757983] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:41.610 [2024-11-19 15:22:31.758022] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:41.610 [2024-11-19 15:22:31.758029] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:41.610 15:22:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:41.610 15:22:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:16:41.610 15:22:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:41.610 15:22:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:41.610 15:22:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:16:41.610 15:22:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:16:41.610 15:22:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:16:41.610 15:22:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:41.610 15:22:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:41.610 15:22:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:41.610 15:22:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:41.610 15:22:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:41.610 15:22:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:41.610 15:22:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:41.610 15:22:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:41.610 15:22:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:41.610 15:22:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:41.610 "name": "raid_bdev1", 00:16:41.610 "uuid": "49213557-9482-4923-a3f8-3e1a78fbbc90", 00:16:41.610 "strip_size_kb": 64, 00:16:41.610 "state": "online", 00:16:41.610 "raid_level": "raid5f", 00:16:41.610 "superblock": true, 00:16:41.610 "num_base_bdevs": 4, 00:16:41.610 "num_base_bdevs_discovered": 3, 00:16:41.610 "num_base_bdevs_operational": 3, 00:16:41.610 "base_bdevs_list": [ 00:16:41.610 { 00:16:41.610 "name": null, 00:16:41.610 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:41.610 "is_configured": false, 00:16:41.610 "data_offset": 0, 00:16:41.610 "data_size": 63488 00:16:41.611 }, 00:16:41.611 { 00:16:41.611 "name": "BaseBdev2", 00:16:41.611 "uuid": "369db297-95ae-56bd-88b3-690234fd2e40", 00:16:41.611 "is_configured": true, 00:16:41.611 "data_offset": 2048, 00:16:41.611 "data_size": 63488 00:16:41.611 }, 00:16:41.611 { 00:16:41.611 "name": "BaseBdev3", 00:16:41.611 "uuid": "d8f88c70-5105-55ff-b37b-778fcf5056bb", 00:16:41.611 "is_configured": true, 00:16:41.611 "data_offset": 2048, 00:16:41.611 "data_size": 63488 00:16:41.611 }, 00:16:41.611 { 00:16:41.611 "name": "BaseBdev4", 00:16:41.611 "uuid": "c1bdaaee-4f20-5259-9a48-adf4beab217a", 00:16:41.611 "is_configured": true, 00:16:41.611 "data_offset": 2048, 00:16:41.611 "data_size": 63488 00:16:41.611 } 00:16:41.611 ] 00:16:41.611 }' 00:16:41.611 15:22:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:41.611 15:22:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:42.180 15:22:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:42.180 15:22:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:42.180 15:22:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:42.180 [2024-11-19 15:22:32.266332] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:42.180 [2024-11-19 15:22:32.266443] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:42.180 [2024-11-19 15:22:32.266486] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b780 00:16:42.180 [2024-11-19 15:22:32.266515] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:42.180 [2024-11-19 15:22:32.266962] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:42.180 [2024-11-19 15:22:32.267038] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:42.180 [2024-11-19 15:22:32.267157] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:16:42.180 [2024-11-19 15:22:32.267196] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:16:42.180 [2024-11-19 15:22:32.267240] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:16:42.180 [2024-11-19 15:22:32.267317] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:42.180 [2024-11-19 15:22:32.271062] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000459c0 00:16:42.180 spare 00:16:42.180 15:22:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:42.180 15:22:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:16:42.180 [2024-11-19 15:22:32.273296] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:43.120 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:43.120 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:43.120 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:43.120 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:43.120 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:43.120 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:43.120 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:43.120 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:43.120 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:43.120 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:43.120 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:43.120 "name": "raid_bdev1", 00:16:43.120 "uuid": "49213557-9482-4923-a3f8-3e1a78fbbc90", 00:16:43.120 "strip_size_kb": 64, 00:16:43.120 "state": "online", 00:16:43.120 "raid_level": "raid5f", 00:16:43.120 "superblock": true, 00:16:43.120 "num_base_bdevs": 4, 00:16:43.120 "num_base_bdevs_discovered": 4, 00:16:43.120 "num_base_bdevs_operational": 4, 00:16:43.120 "process": { 00:16:43.120 "type": "rebuild", 00:16:43.120 "target": "spare", 00:16:43.120 "progress": { 00:16:43.120 "blocks": 19200, 00:16:43.120 "percent": 10 00:16:43.120 } 00:16:43.120 }, 00:16:43.120 "base_bdevs_list": [ 00:16:43.120 { 00:16:43.120 "name": "spare", 00:16:43.120 "uuid": "e6ab5de9-b8b4-50cf-abfb-aa54d1ea5673", 00:16:43.120 "is_configured": true, 00:16:43.120 "data_offset": 2048, 00:16:43.120 "data_size": 63488 00:16:43.120 }, 00:16:43.120 { 00:16:43.120 "name": "BaseBdev2", 00:16:43.120 "uuid": "369db297-95ae-56bd-88b3-690234fd2e40", 00:16:43.120 "is_configured": true, 00:16:43.120 "data_offset": 2048, 00:16:43.120 "data_size": 63488 00:16:43.120 }, 00:16:43.120 { 00:16:43.120 "name": "BaseBdev3", 00:16:43.120 "uuid": "d8f88c70-5105-55ff-b37b-778fcf5056bb", 00:16:43.120 "is_configured": true, 00:16:43.120 "data_offset": 2048, 00:16:43.120 "data_size": 63488 00:16:43.120 }, 00:16:43.120 { 00:16:43.120 "name": "BaseBdev4", 00:16:43.120 "uuid": "c1bdaaee-4f20-5259-9a48-adf4beab217a", 00:16:43.120 "is_configured": true, 00:16:43.120 "data_offset": 2048, 00:16:43.120 "data_size": 63488 00:16:43.120 } 00:16:43.120 ] 00:16:43.120 }' 00:16:43.120 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:43.120 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:43.120 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:43.120 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:43.120 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:16:43.120 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:43.120 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:43.120 [2024-11-19 15:22:33.421192] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:43.393 [2024-11-19 15:22:33.478232] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:43.393 [2024-11-19 15:22:33.478287] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:43.393 [2024-11-19 15:22:33.478302] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:43.393 [2024-11-19 15:22:33.478310] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:43.393 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:43.393 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:16:43.393 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:43.393 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:43.393 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:16:43.393 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:16:43.393 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:16:43.393 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:43.393 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:43.393 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:43.393 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:43.393 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:43.393 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:43.393 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:43.393 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:43.393 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:43.393 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:43.393 "name": "raid_bdev1", 00:16:43.393 "uuid": "49213557-9482-4923-a3f8-3e1a78fbbc90", 00:16:43.393 "strip_size_kb": 64, 00:16:43.393 "state": "online", 00:16:43.393 "raid_level": "raid5f", 00:16:43.393 "superblock": true, 00:16:43.393 "num_base_bdevs": 4, 00:16:43.393 "num_base_bdevs_discovered": 3, 00:16:43.393 "num_base_bdevs_operational": 3, 00:16:43.393 "base_bdevs_list": [ 00:16:43.393 { 00:16:43.393 "name": null, 00:16:43.393 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:43.393 "is_configured": false, 00:16:43.393 "data_offset": 0, 00:16:43.393 "data_size": 63488 00:16:43.393 }, 00:16:43.393 { 00:16:43.393 "name": "BaseBdev2", 00:16:43.393 "uuid": "369db297-95ae-56bd-88b3-690234fd2e40", 00:16:43.393 "is_configured": true, 00:16:43.393 "data_offset": 2048, 00:16:43.393 "data_size": 63488 00:16:43.393 }, 00:16:43.393 { 00:16:43.393 "name": "BaseBdev3", 00:16:43.393 "uuid": "d8f88c70-5105-55ff-b37b-778fcf5056bb", 00:16:43.393 "is_configured": true, 00:16:43.393 "data_offset": 2048, 00:16:43.393 "data_size": 63488 00:16:43.393 }, 00:16:43.393 { 00:16:43.393 "name": "BaseBdev4", 00:16:43.393 "uuid": "c1bdaaee-4f20-5259-9a48-adf4beab217a", 00:16:43.393 "is_configured": true, 00:16:43.393 "data_offset": 2048, 00:16:43.393 "data_size": 63488 00:16:43.393 } 00:16:43.393 ] 00:16:43.393 }' 00:16:43.393 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:43.393 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:43.677 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:43.677 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:43.677 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:43.677 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:43.677 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:43.677 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:43.677 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:43.677 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:43.677 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:43.677 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:43.677 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:43.677 "name": "raid_bdev1", 00:16:43.677 "uuid": "49213557-9482-4923-a3f8-3e1a78fbbc90", 00:16:43.677 "strip_size_kb": 64, 00:16:43.677 "state": "online", 00:16:43.677 "raid_level": "raid5f", 00:16:43.677 "superblock": true, 00:16:43.677 "num_base_bdevs": 4, 00:16:43.677 "num_base_bdevs_discovered": 3, 00:16:43.677 "num_base_bdevs_operational": 3, 00:16:43.677 "base_bdevs_list": [ 00:16:43.677 { 00:16:43.677 "name": null, 00:16:43.677 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:43.677 "is_configured": false, 00:16:43.677 "data_offset": 0, 00:16:43.677 "data_size": 63488 00:16:43.677 }, 00:16:43.677 { 00:16:43.677 "name": "BaseBdev2", 00:16:43.677 "uuid": "369db297-95ae-56bd-88b3-690234fd2e40", 00:16:43.677 "is_configured": true, 00:16:43.677 "data_offset": 2048, 00:16:43.677 "data_size": 63488 00:16:43.677 }, 00:16:43.677 { 00:16:43.677 "name": "BaseBdev3", 00:16:43.677 "uuid": "d8f88c70-5105-55ff-b37b-778fcf5056bb", 00:16:43.677 "is_configured": true, 00:16:43.677 "data_offset": 2048, 00:16:43.677 "data_size": 63488 00:16:43.677 }, 00:16:43.677 { 00:16:43.677 "name": "BaseBdev4", 00:16:43.677 "uuid": "c1bdaaee-4f20-5259-9a48-adf4beab217a", 00:16:43.677 "is_configured": true, 00:16:43.677 "data_offset": 2048, 00:16:43.677 "data_size": 63488 00:16:43.677 } 00:16:43.677 ] 00:16:43.677 }' 00:16:43.677 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:43.677 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:43.677 15:22:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:43.957 15:22:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:43.957 15:22:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:16:43.957 15:22:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:43.957 15:22:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:43.957 15:22:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:43.957 15:22:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:16:43.957 15:22:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:43.957 15:22:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:43.957 [2024-11-19 15:22:34.042499] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:16:43.957 [2024-11-19 15:22:34.042609] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:43.957 [2024-11-19 15:22:34.042643] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000bd80 00:16:43.957 [2024-11-19 15:22:34.042673] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:43.957 [2024-11-19 15:22:34.043093] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:43.957 [2024-11-19 15:22:34.043154] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:16:43.957 [2024-11-19 15:22:34.043253] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:16:43.957 [2024-11-19 15:22:34.043298] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:16:43.957 [2024-11-19 15:22:34.043337] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:16:43.957 [2024-11-19 15:22:34.043372] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:16:43.957 BaseBdev1 00:16:43.957 15:22:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:43.957 15:22:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:16:44.913 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:16:44.913 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:44.913 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:44.913 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:16:44.913 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:16:44.913 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:16:44.913 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:44.913 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:44.913 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:44.913 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:44.913 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:44.913 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:44.913 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:44.913 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:44.913 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:44.913 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:44.913 "name": "raid_bdev1", 00:16:44.913 "uuid": "49213557-9482-4923-a3f8-3e1a78fbbc90", 00:16:44.913 "strip_size_kb": 64, 00:16:44.913 "state": "online", 00:16:44.913 "raid_level": "raid5f", 00:16:44.913 "superblock": true, 00:16:44.913 "num_base_bdevs": 4, 00:16:44.913 "num_base_bdevs_discovered": 3, 00:16:44.913 "num_base_bdevs_operational": 3, 00:16:44.913 "base_bdevs_list": [ 00:16:44.913 { 00:16:44.913 "name": null, 00:16:44.913 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:44.913 "is_configured": false, 00:16:44.913 "data_offset": 0, 00:16:44.913 "data_size": 63488 00:16:44.913 }, 00:16:44.913 { 00:16:44.913 "name": "BaseBdev2", 00:16:44.913 "uuid": "369db297-95ae-56bd-88b3-690234fd2e40", 00:16:44.913 "is_configured": true, 00:16:44.913 "data_offset": 2048, 00:16:44.913 "data_size": 63488 00:16:44.913 }, 00:16:44.913 { 00:16:44.913 "name": "BaseBdev3", 00:16:44.913 "uuid": "d8f88c70-5105-55ff-b37b-778fcf5056bb", 00:16:44.913 "is_configured": true, 00:16:44.913 "data_offset": 2048, 00:16:44.913 "data_size": 63488 00:16:44.913 }, 00:16:44.913 { 00:16:44.913 "name": "BaseBdev4", 00:16:44.913 "uuid": "c1bdaaee-4f20-5259-9a48-adf4beab217a", 00:16:44.913 "is_configured": true, 00:16:44.913 "data_offset": 2048, 00:16:44.913 "data_size": 63488 00:16:44.913 } 00:16:44.913 ] 00:16:44.913 }' 00:16:44.913 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:44.913 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:45.173 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:45.173 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:45.173 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:45.173 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:45.173 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:45.173 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:45.433 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:45.433 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:45.433 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:45.433 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:45.433 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:45.433 "name": "raid_bdev1", 00:16:45.433 "uuid": "49213557-9482-4923-a3f8-3e1a78fbbc90", 00:16:45.433 "strip_size_kb": 64, 00:16:45.433 "state": "online", 00:16:45.433 "raid_level": "raid5f", 00:16:45.433 "superblock": true, 00:16:45.433 "num_base_bdevs": 4, 00:16:45.433 "num_base_bdevs_discovered": 3, 00:16:45.433 "num_base_bdevs_operational": 3, 00:16:45.433 "base_bdevs_list": [ 00:16:45.433 { 00:16:45.433 "name": null, 00:16:45.433 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:45.433 "is_configured": false, 00:16:45.433 "data_offset": 0, 00:16:45.433 "data_size": 63488 00:16:45.433 }, 00:16:45.433 { 00:16:45.433 "name": "BaseBdev2", 00:16:45.433 "uuid": "369db297-95ae-56bd-88b3-690234fd2e40", 00:16:45.433 "is_configured": true, 00:16:45.433 "data_offset": 2048, 00:16:45.433 "data_size": 63488 00:16:45.433 }, 00:16:45.433 { 00:16:45.433 "name": "BaseBdev3", 00:16:45.433 "uuid": "d8f88c70-5105-55ff-b37b-778fcf5056bb", 00:16:45.433 "is_configured": true, 00:16:45.433 "data_offset": 2048, 00:16:45.433 "data_size": 63488 00:16:45.433 }, 00:16:45.433 { 00:16:45.433 "name": "BaseBdev4", 00:16:45.433 "uuid": "c1bdaaee-4f20-5259-9a48-adf4beab217a", 00:16:45.433 "is_configured": true, 00:16:45.433 "data_offset": 2048, 00:16:45.433 "data_size": 63488 00:16:45.433 } 00:16:45.433 ] 00:16:45.433 }' 00:16:45.433 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:45.433 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:45.434 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:45.434 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:45.434 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:45.434 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@652 -- # local es=0 00:16:45.434 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:45.434 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:16:45.434 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:16:45.434 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:16:45.434 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:16:45.434 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:45.434 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:45.434 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:45.434 [2024-11-19 15:22:35.659907] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:45.434 [2024-11-19 15:22:35.660103] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:16:45.434 [2024-11-19 15:22:35.660160] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:16:45.434 request: 00:16:45.434 { 00:16:45.434 "base_bdev": "BaseBdev1", 00:16:45.434 "raid_bdev": "raid_bdev1", 00:16:45.434 "method": "bdev_raid_add_base_bdev", 00:16:45.434 "req_id": 1 00:16:45.434 } 00:16:45.434 Got JSON-RPC error response 00:16:45.434 response: 00:16:45.434 { 00:16:45.434 "code": -22, 00:16:45.434 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:16:45.434 } 00:16:45.434 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:16:45.434 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@655 -- # es=1 00:16:45.434 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:16:45.434 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:16:45.434 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:16:45.434 15:22:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:16:46.375 15:22:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:16:46.375 15:22:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:46.375 15:22:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:46.375 15:22:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:16:46.375 15:22:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:16:46.375 15:22:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:16:46.375 15:22:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:46.375 15:22:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:46.375 15:22:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:46.375 15:22:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:46.375 15:22:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:46.375 15:22:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:46.376 15:22:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:46.376 15:22:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:46.376 15:22:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:46.636 15:22:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:46.636 "name": "raid_bdev1", 00:16:46.636 "uuid": "49213557-9482-4923-a3f8-3e1a78fbbc90", 00:16:46.636 "strip_size_kb": 64, 00:16:46.636 "state": "online", 00:16:46.636 "raid_level": "raid5f", 00:16:46.636 "superblock": true, 00:16:46.636 "num_base_bdevs": 4, 00:16:46.636 "num_base_bdevs_discovered": 3, 00:16:46.636 "num_base_bdevs_operational": 3, 00:16:46.636 "base_bdevs_list": [ 00:16:46.636 { 00:16:46.636 "name": null, 00:16:46.636 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:46.636 "is_configured": false, 00:16:46.636 "data_offset": 0, 00:16:46.636 "data_size": 63488 00:16:46.636 }, 00:16:46.636 { 00:16:46.636 "name": "BaseBdev2", 00:16:46.636 "uuid": "369db297-95ae-56bd-88b3-690234fd2e40", 00:16:46.636 "is_configured": true, 00:16:46.636 "data_offset": 2048, 00:16:46.636 "data_size": 63488 00:16:46.636 }, 00:16:46.636 { 00:16:46.636 "name": "BaseBdev3", 00:16:46.636 "uuid": "d8f88c70-5105-55ff-b37b-778fcf5056bb", 00:16:46.636 "is_configured": true, 00:16:46.636 "data_offset": 2048, 00:16:46.636 "data_size": 63488 00:16:46.636 }, 00:16:46.636 { 00:16:46.636 "name": "BaseBdev4", 00:16:46.636 "uuid": "c1bdaaee-4f20-5259-9a48-adf4beab217a", 00:16:46.636 "is_configured": true, 00:16:46.636 "data_offset": 2048, 00:16:46.636 "data_size": 63488 00:16:46.636 } 00:16:46.636 ] 00:16:46.636 }' 00:16:46.636 15:22:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:46.636 15:22:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:46.896 15:22:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:46.896 15:22:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:46.896 15:22:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:46.896 15:22:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:46.896 15:22:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:46.896 15:22:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:46.896 15:22:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:46.896 15:22:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:46.896 15:22:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:46.896 15:22:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:46.896 15:22:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:46.896 "name": "raid_bdev1", 00:16:46.896 "uuid": "49213557-9482-4923-a3f8-3e1a78fbbc90", 00:16:46.896 "strip_size_kb": 64, 00:16:46.896 "state": "online", 00:16:46.896 "raid_level": "raid5f", 00:16:46.896 "superblock": true, 00:16:46.896 "num_base_bdevs": 4, 00:16:46.896 "num_base_bdevs_discovered": 3, 00:16:46.896 "num_base_bdevs_operational": 3, 00:16:46.896 "base_bdevs_list": [ 00:16:46.896 { 00:16:46.896 "name": null, 00:16:46.896 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:46.896 "is_configured": false, 00:16:46.896 "data_offset": 0, 00:16:46.896 "data_size": 63488 00:16:46.896 }, 00:16:46.896 { 00:16:46.896 "name": "BaseBdev2", 00:16:46.896 "uuid": "369db297-95ae-56bd-88b3-690234fd2e40", 00:16:46.896 "is_configured": true, 00:16:46.896 "data_offset": 2048, 00:16:46.896 "data_size": 63488 00:16:46.896 }, 00:16:46.896 { 00:16:46.896 "name": "BaseBdev3", 00:16:46.896 "uuid": "d8f88c70-5105-55ff-b37b-778fcf5056bb", 00:16:46.896 "is_configured": true, 00:16:46.896 "data_offset": 2048, 00:16:46.896 "data_size": 63488 00:16:46.896 }, 00:16:46.896 { 00:16:46.896 "name": "BaseBdev4", 00:16:46.896 "uuid": "c1bdaaee-4f20-5259-9a48-adf4beab217a", 00:16:46.896 "is_configured": true, 00:16:46.896 "data_offset": 2048, 00:16:46.896 "data_size": 63488 00:16:46.896 } 00:16:46.896 ] 00:16:46.896 }' 00:16:46.896 15:22:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:46.896 15:22:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:46.896 15:22:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:47.157 15:22:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:47.157 15:22:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 95530 00:16:47.157 15:22:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@954 -- # '[' -z 95530 ']' 00:16:47.157 15:22:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@958 -- # kill -0 95530 00:16:47.157 15:22:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@959 -- # uname 00:16:47.157 15:22:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:16:47.157 15:22:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 95530 00:16:47.157 15:22:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:16:47.157 15:22:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:16:47.157 15:22:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 95530' 00:16:47.157 killing process with pid 95530 00:16:47.157 15:22:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@973 -- # kill 95530 00:16:47.157 Received shutdown signal, test time was about 60.000000 seconds 00:16:47.157 00:16:47.157 Latency(us) 00:16:47.157 [2024-11-19T15:22:37.496Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:16:47.157 [2024-11-19T15:22:37.496Z] =================================================================================================================== 00:16:47.157 [2024-11-19T15:22:37.496Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:16:47.157 [2024-11-19 15:22:37.312646] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:47.157 [2024-11-19 15:22:37.312748] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:47.157 [2024-11-19 15:22:37.312818] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:47.157 [2024-11-19 15:22:37.312827] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:16:47.157 15:22:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@978 -- # wait 95530 00:16:47.157 [2024-11-19 15:22:37.364023] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:47.418 15:22:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:16:47.418 00:16:47.418 real 0m25.214s 00:16:47.418 user 0m32.053s 00:16:47.418 sys 0m3.087s 00:16:47.418 15:22:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:47.418 15:22:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:16:47.418 ************************************ 00:16:47.418 END TEST raid5f_rebuild_test_sb 00:16:47.418 ************************************ 00:16:47.418 15:22:37 bdev_raid -- bdev/bdev_raid.sh@995 -- # base_blocklen=4096 00:16:47.418 15:22:37 bdev_raid -- bdev/bdev_raid.sh@997 -- # run_test raid_state_function_test_sb_4k raid_state_function_test raid1 2 true 00:16:47.418 15:22:37 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:16:47.418 15:22:37 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:47.418 15:22:37 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:47.418 ************************************ 00:16:47.418 START TEST raid_state_function_test_sb_4k 00:16:47.418 ************************************ 00:16:47.418 15:22:37 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 2 true 00:16:47.418 15:22:37 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:16:47.418 15:22:37 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:16:47.418 15:22:37 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:16:47.418 15:22:37 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:16:47.418 15:22:37 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:16:47.418 15:22:37 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:16:47.418 15:22:37 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:16:47.418 15:22:37 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:16:47.418 15:22:37 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:16:47.418 15:22:37 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:16:47.418 15:22:37 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:16:47.418 15:22:37 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:16:47.418 15:22:37 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:16:47.418 15:22:37 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:16:47.418 15:22:37 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:16:47.418 15:22:37 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@211 -- # local strip_size 00:16:47.418 15:22:37 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:16:47.418 15:22:37 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:16:47.418 15:22:37 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:16:47.418 15:22:37 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:16:47.418 15:22:37 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:16:47.418 15:22:37 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:16:47.418 15:22:37 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@229 -- # raid_pid=96324 00:16:47.418 15:22:37 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:16:47.418 Process raid pid: 96324 00:16:47.418 15:22:37 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 96324' 00:16:47.418 15:22:37 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@231 -- # waitforlisten 96324 00:16:47.418 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:47.418 15:22:37 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@835 -- # '[' -z 96324 ']' 00:16:47.418 15:22:37 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:47.418 15:22:37 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@840 -- # local max_retries=100 00:16:47.418 15:22:37 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:47.418 15:22:37 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@844 -- # xtrace_disable 00:16:47.418 15:22:37 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:47.418 [2024-11-19 15:22:37.736196] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:16:47.418 [2024-11-19 15:22:37.736317] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:16:47.678 [2024-11-19 15:22:37.893403] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:47.678 [2024-11-19 15:22:37.917993] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:47.678 [2024-11-19 15:22:37.961062] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:47.678 [2024-11-19 15:22:37.961100] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:48.249 15:22:38 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:16:48.249 15:22:38 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@868 -- # return 0 00:16:48.249 15:22:38 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:16:48.249 15:22:38 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:48.249 15:22:38 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:48.249 [2024-11-19 15:22:38.558760] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:16:48.249 [2024-11-19 15:22:38.558818] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:16:48.249 [2024-11-19 15:22:38.558828] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:48.249 [2024-11-19 15:22:38.558838] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:48.249 15:22:38 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:48.249 15:22:38 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:16:48.249 15:22:38 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:48.249 15:22:38 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:48.249 15:22:38 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:48.249 15:22:38 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:48.249 15:22:38 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:48.249 15:22:38 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:48.249 15:22:38 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:48.249 15:22:38 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:48.249 15:22:38 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:48.249 15:22:38 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:48.249 15:22:38 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:48.249 15:22:38 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:48.249 15:22:38 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:48.509 15:22:38 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:48.509 15:22:38 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:48.509 "name": "Existed_Raid", 00:16:48.509 "uuid": "b91be37e-11ef-49e5-b0eb-77ebebcc9b8f", 00:16:48.509 "strip_size_kb": 0, 00:16:48.509 "state": "configuring", 00:16:48.509 "raid_level": "raid1", 00:16:48.509 "superblock": true, 00:16:48.509 "num_base_bdevs": 2, 00:16:48.509 "num_base_bdevs_discovered": 0, 00:16:48.509 "num_base_bdevs_operational": 2, 00:16:48.509 "base_bdevs_list": [ 00:16:48.509 { 00:16:48.509 "name": "BaseBdev1", 00:16:48.509 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:48.509 "is_configured": false, 00:16:48.509 "data_offset": 0, 00:16:48.509 "data_size": 0 00:16:48.509 }, 00:16:48.509 { 00:16:48.509 "name": "BaseBdev2", 00:16:48.509 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:48.509 "is_configured": false, 00:16:48.509 "data_offset": 0, 00:16:48.509 "data_size": 0 00:16:48.509 } 00:16:48.509 ] 00:16:48.509 }' 00:16:48.509 15:22:38 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:48.509 15:22:38 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:48.770 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:16:48.770 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:48.770 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:48.770 [2024-11-19 15:22:39.033918] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:16:48.770 [2024-11-19 15:22:39.034033] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:16:48.770 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:48.770 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:16:48.770 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:48.770 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:48.770 [2024-11-19 15:22:39.045915] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:16:48.770 [2024-11-19 15:22:39.046021] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:16:48.770 [2024-11-19 15:22:39.046050] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:48.770 [2024-11-19 15:22:39.046085] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:48.770 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:48.770 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev1 00:16:48.770 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:48.770 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:48.770 [2024-11-19 15:22:39.066810] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:48.770 BaseBdev1 00:16:48.770 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:48.770 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:16:48.770 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:16:48.770 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:16:48.770 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@905 -- # local i 00:16:48.770 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:16:48.770 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:16:48.770 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:16:48.770 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:48.770 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:48.770 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:48.770 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:16:48.770 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:48.770 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:48.770 [ 00:16:48.770 { 00:16:48.770 "name": "BaseBdev1", 00:16:48.770 "aliases": [ 00:16:48.770 "0fd682c4-9fc1-499f-9d9c-4a835cd6d409" 00:16:48.770 ], 00:16:48.770 "product_name": "Malloc disk", 00:16:48.770 "block_size": 4096, 00:16:48.770 "num_blocks": 8192, 00:16:48.770 "uuid": "0fd682c4-9fc1-499f-9d9c-4a835cd6d409", 00:16:48.770 "assigned_rate_limits": { 00:16:48.770 "rw_ios_per_sec": 0, 00:16:48.770 "rw_mbytes_per_sec": 0, 00:16:48.770 "r_mbytes_per_sec": 0, 00:16:48.770 "w_mbytes_per_sec": 0 00:16:48.770 }, 00:16:48.770 "claimed": true, 00:16:48.770 "claim_type": "exclusive_write", 00:16:48.770 "zoned": false, 00:16:48.770 "supported_io_types": { 00:16:48.770 "read": true, 00:16:48.770 "write": true, 00:16:48.770 "unmap": true, 00:16:48.770 "flush": true, 00:16:48.770 "reset": true, 00:16:48.770 "nvme_admin": false, 00:16:48.770 "nvme_io": false, 00:16:48.770 "nvme_io_md": false, 00:16:48.770 "write_zeroes": true, 00:16:48.770 "zcopy": true, 00:16:48.770 "get_zone_info": false, 00:16:48.770 "zone_management": false, 00:16:48.770 "zone_append": false, 00:16:48.770 "compare": false, 00:16:48.770 "compare_and_write": false, 00:16:48.770 "abort": true, 00:16:48.770 "seek_hole": false, 00:16:48.770 "seek_data": false, 00:16:48.770 "copy": true, 00:16:48.770 "nvme_iov_md": false 00:16:48.770 }, 00:16:48.770 "memory_domains": [ 00:16:48.770 { 00:16:48.770 "dma_device_id": "system", 00:16:48.770 "dma_device_type": 1 00:16:48.770 }, 00:16:48.770 { 00:16:49.039 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:49.039 "dma_device_type": 2 00:16:49.039 } 00:16:49.039 ], 00:16:49.039 "driver_specific": {} 00:16:49.039 } 00:16:49.039 ] 00:16:49.039 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:49.039 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@911 -- # return 0 00:16:49.039 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:16:49.039 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:49.039 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:49.039 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:49.039 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:49.039 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:49.039 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:49.039 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:49.039 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:49.039 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:49.039 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:49.039 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:49.039 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:49.039 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:49.039 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:49.039 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:49.039 "name": "Existed_Raid", 00:16:49.039 "uuid": "119adc7c-96e0-4d6c-958f-b9f7ed34f189", 00:16:49.039 "strip_size_kb": 0, 00:16:49.039 "state": "configuring", 00:16:49.039 "raid_level": "raid1", 00:16:49.039 "superblock": true, 00:16:49.039 "num_base_bdevs": 2, 00:16:49.039 "num_base_bdevs_discovered": 1, 00:16:49.039 "num_base_bdevs_operational": 2, 00:16:49.039 "base_bdevs_list": [ 00:16:49.039 { 00:16:49.039 "name": "BaseBdev1", 00:16:49.039 "uuid": "0fd682c4-9fc1-499f-9d9c-4a835cd6d409", 00:16:49.039 "is_configured": true, 00:16:49.039 "data_offset": 256, 00:16:49.039 "data_size": 7936 00:16:49.039 }, 00:16:49.039 { 00:16:49.039 "name": "BaseBdev2", 00:16:49.039 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:49.039 "is_configured": false, 00:16:49.039 "data_offset": 0, 00:16:49.039 "data_size": 0 00:16:49.039 } 00:16:49.040 ] 00:16:49.040 }' 00:16:49.040 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:49.040 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:49.309 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:16:49.309 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:49.309 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:49.309 [2024-11-19 15:22:39.546008] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:16:49.309 [2024-11-19 15:22:39.546101] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:16:49.309 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:49.309 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:16:49.309 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:49.309 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:49.309 [2024-11-19 15:22:39.558031] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:49.309 [2024-11-19 15:22:39.559838] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:49.309 [2024-11-19 15:22:39.559880] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:49.309 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:49.309 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:16:49.309 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:16:49.309 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:16:49.309 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:49.309 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:49.309 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:49.309 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:49.309 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:49.309 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:49.309 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:49.309 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:49.309 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:49.309 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:49.309 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:49.309 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:49.309 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:49.309 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:49.309 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:49.309 "name": "Existed_Raid", 00:16:49.309 "uuid": "54480bad-8de4-442f-95c5-c0fb236580a2", 00:16:49.309 "strip_size_kb": 0, 00:16:49.309 "state": "configuring", 00:16:49.309 "raid_level": "raid1", 00:16:49.309 "superblock": true, 00:16:49.309 "num_base_bdevs": 2, 00:16:49.309 "num_base_bdevs_discovered": 1, 00:16:49.309 "num_base_bdevs_operational": 2, 00:16:49.309 "base_bdevs_list": [ 00:16:49.309 { 00:16:49.309 "name": "BaseBdev1", 00:16:49.309 "uuid": "0fd682c4-9fc1-499f-9d9c-4a835cd6d409", 00:16:49.309 "is_configured": true, 00:16:49.309 "data_offset": 256, 00:16:49.309 "data_size": 7936 00:16:49.309 }, 00:16:49.309 { 00:16:49.309 "name": "BaseBdev2", 00:16:49.309 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:49.309 "is_configured": false, 00:16:49.309 "data_offset": 0, 00:16:49.309 "data_size": 0 00:16:49.309 } 00:16:49.309 ] 00:16:49.309 }' 00:16:49.309 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:49.309 15:22:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:49.879 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev2 00:16:49.880 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:49.880 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:49.880 [2024-11-19 15:22:40.024100] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:49.880 [2024-11-19 15:22:40.024380] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:16:49.880 [2024-11-19 15:22:40.024439] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:49.880 [2024-11-19 15:22:40.024721] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:16:49.880 BaseBdev2 00:16:49.880 [2024-11-19 15:22:40.024909] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:16:49.880 [2024-11-19 15:22:40.024930] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:16:49.880 [2024-11-19 15:22:40.025057] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:49.880 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:49.880 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:16:49.880 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:16:49.880 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:16:49.880 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@905 -- # local i 00:16:49.880 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:16:49.880 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:16:49.880 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:16:49.880 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:49.880 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:49.880 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:49.880 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:16:49.880 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:49.880 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:49.880 [ 00:16:49.880 { 00:16:49.880 "name": "BaseBdev2", 00:16:49.880 "aliases": [ 00:16:49.880 "53257345-7102-438e-b7dd-bf0b5bd154c7" 00:16:49.880 ], 00:16:49.880 "product_name": "Malloc disk", 00:16:49.880 "block_size": 4096, 00:16:49.880 "num_blocks": 8192, 00:16:49.880 "uuid": "53257345-7102-438e-b7dd-bf0b5bd154c7", 00:16:49.880 "assigned_rate_limits": { 00:16:49.880 "rw_ios_per_sec": 0, 00:16:49.880 "rw_mbytes_per_sec": 0, 00:16:49.880 "r_mbytes_per_sec": 0, 00:16:49.880 "w_mbytes_per_sec": 0 00:16:49.880 }, 00:16:49.880 "claimed": true, 00:16:49.880 "claim_type": "exclusive_write", 00:16:49.880 "zoned": false, 00:16:49.880 "supported_io_types": { 00:16:49.880 "read": true, 00:16:49.880 "write": true, 00:16:49.880 "unmap": true, 00:16:49.880 "flush": true, 00:16:49.880 "reset": true, 00:16:49.880 "nvme_admin": false, 00:16:49.880 "nvme_io": false, 00:16:49.880 "nvme_io_md": false, 00:16:49.880 "write_zeroes": true, 00:16:49.880 "zcopy": true, 00:16:49.880 "get_zone_info": false, 00:16:49.880 "zone_management": false, 00:16:49.880 "zone_append": false, 00:16:49.880 "compare": false, 00:16:49.880 "compare_and_write": false, 00:16:49.880 "abort": true, 00:16:49.880 "seek_hole": false, 00:16:49.880 "seek_data": false, 00:16:49.880 "copy": true, 00:16:49.880 "nvme_iov_md": false 00:16:49.880 }, 00:16:49.880 "memory_domains": [ 00:16:49.880 { 00:16:49.880 "dma_device_id": "system", 00:16:49.880 "dma_device_type": 1 00:16:49.880 }, 00:16:49.880 { 00:16:49.880 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:49.880 "dma_device_type": 2 00:16:49.880 } 00:16:49.880 ], 00:16:49.880 "driver_specific": {} 00:16:49.880 } 00:16:49.880 ] 00:16:49.880 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:49.880 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@911 -- # return 0 00:16:49.880 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:16:49.880 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:16:49.880 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:16:49.880 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:49.880 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:49.880 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:49.880 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:49.880 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:49.880 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:49.880 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:49.880 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:49.880 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:49.880 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:49.880 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:49.880 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:49.880 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:49.880 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:49.880 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:49.880 "name": "Existed_Raid", 00:16:49.880 "uuid": "54480bad-8de4-442f-95c5-c0fb236580a2", 00:16:49.880 "strip_size_kb": 0, 00:16:49.880 "state": "online", 00:16:49.880 "raid_level": "raid1", 00:16:49.880 "superblock": true, 00:16:49.880 "num_base_bdevs": 2, 00:16:49.880 "num_base_bdevs_discovered": 2, 00:16:49.880 "num_base_bdevs_operational": 2, 00:16:49.880 "base_bdevs_list": [ 00:16:49.880 { 00:16:49.880 "name": "BaseBdev1", 00:16:49.880 "uuid": "0fd682c4-9fc1-499f-9d9c-4a835cd6d409", 00:16:49.880 "is_configured": true, 00:16:49.880 "data_offset": 256, 00:16:49.880 "data_size": 7936 00:16:49.880 }, 00:16:49.880 { 00:16:49.880 "name": "BaseBdev2", 00:16:49.880 "uuid": "53257345-7102-438e-b7dd-bf0b5bd154c7", 00:16:49.880 "is_configured": true, 00:16:49.880 "data_offset": 256, 00:16:49.880 "data_size": 7936 00:16:49.880 } 00:16:49.880 ] 00:16:49.880 }' 00:16:49.880 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:49.880 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:50.451 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:16:50.451 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:16:50.451 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:16:50.451 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:16:50.451 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@184 -- # local name 00:16:50.451 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:16:50.451 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:16:50.451 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:16:50.451 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:50.451 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:50.451 [2024-11-19 15:22:40.519941] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:50.451 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:50.451 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:16:50.451 "name": "Existed_Raid", 00:16:50.451 "aliases": [ 00:16:50.451 "54480bad-8de4-442f-95c5-c0fb236580a2" 00:16:50.451 ], 00:16:50.451 "product_name": "Raid Volume", 00:16:50.451 "block_size": 4096, 00:16:50.451 "num_blocks": 7936, 00:16:50.451 "uuid": "54480bad-8de4-442f-95c5-c0fb236580a2", 00:16:50.451 "assigned_rate_limits": { 00:16:50.451 "rw_ios_per_sec": 0, 00:16:50.451 "rw_mbytes_per_sec": 0, 00:16:50.451 "r_mbytes_per_sec": 0, 00:16:50.451 "w_mbytes_per_sec": 0 00:16:50.451 }, 00:16:50.451 "claimed": false, 00:16:50.451 "zoned": false, 00:16:50.451 "supported_io_types": { 00:16:50.451 "read": true, 00:16:50.452 "write": true, 00:16:50.452 "unmap": false, 00:16:50.452 "flush": false, 00:16:50.452 "reset": true, 00:16:50.452 "nvme_admin": false, 00:16:50.452 "nvme_io": false, 00:16:50.452 "nvme_io_md": false, 00:16:50.452 "write_zeroes": true, 00:16:50.452 "zcopy": false, 00:16:50.452 "get_zone_info": false, 00:16:50.452 "zone_management": false, 00:16:50.452 "zone_append": false, 00:16:50.452 "compare": false, 00:16:50.452 "compare_and_write": false, 00:16:50.452 "abort": false, 00:16:50.452 "seek_hole": false, 00:16:50.452 "seek_data": false, 00:16:50.452 "copy": false, 00:16:50.452 "nvme_iov_md": false 00:16:50.452 }, 00:16:50.452 "memory_domains": [ 00:16:50.452 { 00:16:50.452 "dma_device_id": "system", 00:16:50.452 "dma_device_type": 1 00:16:50.452 }, 00:16:50.452 { 00:16:50.452 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:50.452 "dma_device_type": 2 00:16:50.452 }, 00:16:50.452 { 00:16:50.452 "dma_device_id": "system", 00:16:50.452 "dma_device_type": 1 00:16:50.452 }, 00:16:50.452 { 00:16:50.452 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:50.452 "dma_device_type": 2 00:16:50.452 } 00:16:50.452 ], 00:16:50.452 "driver_specific": { 00:16:50.452 "raid": { 00:16:50.452 "uuid": "54480bad-8de4-442f-95c5-c0fb236580a2", 00:16:50.452 "strip_size_kb": 0, 00:16:50.452 "state": "online", 00:16:50.452 "raid_level": "raid1", 00:16:50.452 "superblock": true, 00:16:50.452 "num_base_bdevs": 2, 00:16:50.452 "num_base_bdevs_discovered": 2, 00:16:50.452 "num_base_bdevs_operational": 2, 00:16:50.452 "base_bdevs_list": [ 00:16:50.452 { 00:16:50.452 "name": "BaseBdev1", 00:16:50.452 "uuid": "0fd682c4-9fc1-499f-9d9c-4a835cd6d409", 00:16:50.452 "is_configured": true, 00:16:50.452 "data_offset": 256, 00:16:50.452 "data_size": 7936 00:16:50.452 }, 00:16:50.452 { 00:16:50.452 "name": "BaseBdev2", 00:16:50.452 "uuid": "53257345-7102-438e-b7dd-bf0b5bd154c7", 00:16:50.452 "is_configured": true, 00:16:50.452 "data_offset": 256, 00:16:50.452 "data_size": 7936 00:16:50.452 } 00:16:50.452 ] 00:16:50.452 } 00:16:50.452 } 00:16:50.452 }' 00:16:50.452 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:50.452 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:16:50.452 BaseBdev2' 00:16:50.452 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:50.452 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 ' 00:16:50.452 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:50.452 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:50.452 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:16:50.452 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:50.452 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:50.452 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:50.452 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:16:50.452 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:16:50.452 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:50.452 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:16:50.452 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:50.452 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:50.452 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:50.452 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:50.452 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:16:50.452 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:16:50.452 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:16:50.452 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:50.452 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:50.452 [2024-11-19 15:22:40.735381] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:16:50.452 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:50.452 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@260 -- # local expected_state 00:16:50.452 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:16:50.452 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@198 -- # case $1 in 00:16:50.452 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@199 -- # return 0 00:16:50.452 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:16:50.452 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:16:50.452 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:50.452 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:50.452 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:50.452 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:50.452 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:50.452 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:50.452 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:50.452 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:50.452 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:50.452 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:50.452 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:50.452 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:50.452 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:50.452 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:50.712 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:50.712 "name": "Existed_Raid", 00:16:50.712 "uuid": "54480bad-8de4-442f-95c5-c0fb236580a2", 00:16:50.712 "strip_size_kb": 0, 00:16:50.712 "state": "online", 00:16:50.712 "raid_level": "raid1", 00:16:50.712 "superblock": true, 00:16:50.712 "num_base_bdevs": 2, 00:16:50.712 "num_base_bdevs_discovered": 1, 00:16:50.712 "num_base_bdevs_operational": 1, 00:16:50.712 "base_bdevs_list": [ 00:16:50.712 { 00:16:50.712 "name": null, 00:16:50.712 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:50.712 "is_configured": false, 00:16:50.712 "data_offset": 0, 00:16:50.712 "data_size": 7936 00:16:50.712 }, 00:16:50.712 { 00:16:50.712 "name": "BaseBdev2", 00:16:50.712 "uuid": "53257345-7102-438e-b7dd-bf0b5bd154c7", 00:16:50.712 "is_configured": true, 00:16:50.712 "data_offset": 256, 00:16:50.712 "data_size": 7936 00:16:50.712 } 00:16:50.712 ] 00:16:50.713 }' 00:16:50.713 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:50.713 15:22:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:50.973 15:22:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:16:50.973 15:22:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:16:50.973 15:22:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:50.973 15:22:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:50.973 15:22:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:50.973 15:22:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:16:50.973 15:22:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:50.973 15:22:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:16:50.973 15:22:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:16:50.973 15:22:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:16:50.973 15:22:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:50.973 15:22:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:50.973 [2024-11-19 15:22:41.273616] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:16:50.973 [2024-11-19 15:22:41.273704] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:50.973 [2024-11-19 15:22:41.285201] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:50.973 [2024-11-19 15:22:41.285312] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:50.973 [2024-11-19 15:22:41.285353] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:16:50.973 15:22:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:50.973 15:22:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:16:50.973 15:22:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:16:50.973 15:22:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:50.973 15:22:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:16:50.973 15:22:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:50.973 15:22:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:50.973 15:22:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:51.233 15:22:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:16:51.233 15:22:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:16:51.233 15:22:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:16:51.233 15:22:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@326 -- # killprocess 96324 00:16:51.233 15:22:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@954 -- # '[' -z 96324 ']' 00:16:51.233 15:22:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@958 -- # kill -0 96324 00:16:51.233 15:22:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@959 -- # uname 00:16:51.233 15:22:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:16:51.233 15:22:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 96324 00:16:51.233 killing process with pid 96324 00:16:51.233 15:22:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:16:51.233 15:22:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:16:51.233 15:22:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@972 -- # echo 'killing process with pid 96324' 00:16:51.233 15:22:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@973 -- # kill 96324 00:16:51.233 [2024-11-19 15:22:41.382288] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:51.233 15:22:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@978 -- # wait 96324 00:16:51.233 [2024-11-19 15:22:41.383279] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:51.493 15:22:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@328 -- # return 0 00:16:51.493 00:16:51.493 real 0m3.954s 00:16:51.493 user 0m6.254s 00:16:51.493 sys 0m0.826s 00:16:51.493 15:22:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:51.493 15:22:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:51.493 ************************************ 00:16:51.493 END TEST raid_state_function_test_sb_4k 00:16:51.493 ************************************ 00:16:51.494 15:22:41 bdev_raid -- bdev/bdev_raid.sh@998 -- # run_test raid_superblock_test_4k raid_superblock_test raid1 2 00:16:51.494 15:22:41 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:16:51.494 15:22:41 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:51.494 15:22:41 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:51.494 ************************************ 00:16:51.494 START TEST raid_superblock_test_4k 00:16:51.494 ************************************ 00:16:51.494 15:22:41 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@1129 -- # raid_superblock_test raid1 2 00:16:51.494 15:22:41 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:16:51.494 15:22:41 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:16:51.494 15:22:41 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:16:51.494 15:22:41 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:16:51.494 15:22:41 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:16:51.494 15:22:41 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:16:51.494 15:22:41 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:16:51.494 15:22:41 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:16:51.494 15:22:41 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:16:51.494 15:22:41 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@399 -- # local strip_size 00:16:51.494 15:22:41 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:16:51.494 15:22:41 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:16:51.494 15:22:41 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:16:51.494 15:22:41 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:16:51.494 15:22:41 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:16:51.494 15:22:41 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@412 -- # raid_pid=96564 00:16:51.494 15:22:41 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:16:51.494 15:22:41 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@413 -- # waitforlisten 96564 00:16:51.494 15:22:41 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@835 -- # '[' -z 96564 ']' 00:16:51.494 15:22:41 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:51.494 15:22:41 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@840 -- # local max_retries=100 00:16:51.494 15:22:41 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:51.494 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:51.494 15:22:41 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@844 -- # xtrace_disable 00:16:51.494 15:22:41 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:51.494 [2024-11-19 15:22:41.771226] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:16:51.494 [2024-11-19 15:22:41.771430] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid96564 ] 00:16:51.754 [2024-11-19 15:22:41.928624] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:51.754 [2024-11-19 15:22:41.954356] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:51.754 [2024-11-19 15:22:41.997582] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:51.754 [2024-11-19 15:22:41.997706] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:52.324 15:22:42 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:16:52.324 15:22:42 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@868 -- # return 0 00:16:52.324 15:22:42 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:16:52.324 15:22:42 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:52.324 15:22:42 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:16:52.324 15:22:42 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:16:52.324 15:22:42 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:16:52.325 15:22:42 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:16:52.325 15:22:42 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:16:52.325 15:22:42 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:16:52.325 15:22:42 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -b malloc1 00:16:52.325 15:22:42 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:52.325 15:22:42 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:52.325 malloc1 00:16:52.325 15:22:42 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:52.325 15:22:42 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:52.325 15:22:42 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:52.325 15:22:42 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:52.325 [2024-11-19 15:22:42.616110] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:52.325 [2024-11-19 15:22:42.616165] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:52.325 [2024-11-19 15:22:42.616188] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:16:52.325 [2024-11-19 15:22:42.616203] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:52.325 [2024-11-19 15:22:42.618260] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:52.325 [2024-11-19 15:22:42.618301] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:52.325 pt1 00:16:52.325 15:22:42 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:52.325 15:22:42 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:16:52.325 15:22:42 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:52.325 15:22:42 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:16:52.325 15:22:42 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:16:52.325 15:22:42 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:16:52.325 15:22:42 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:16:52.325 15:22:42 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:16:52.325 15:22:42 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:16:52.325 15:22:42 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -b malloc2 00:16:52.325 15:22:42 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:52.325 15:22:42 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:52.325 malloc2 00:16:52.325 15:22:42 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:52.325 15:22:42 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:52.325 15:22:42 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:52.325 15:22:42 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:52.325 [2024-11-19 15:22:42.644574] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:52.325 [2024-11-19 15:22:42.644662] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:52.325 [2024-11-19 15:22:42.644693] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:16:52.325 [2024-11-19 15:22:42.644721] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:52.325 [2024-11-19 15:22:42.646748] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:52.325 [2024-11-19 15:22:42.646817] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:52.325 pt2 00:16:52.325 15:22:42 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:52.325 15:22:42 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:16:52.325 15:22:42 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:52.325 15:22:42 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:16:52.325 15:22:42 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:52.325 15:22:42 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:52.325 [2024-11-19 15:22:42.656591] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:52.325 [2024-11-19 15:22:42.658429] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:52.325 [2024-11-19 15:22:42.658616] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:16:52.325 [2024-11-19 15:22:42.658665] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:52.325 [2024-11-19 15:22:42.658954] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:16:52.325 [2024-11-19 15:22:42.659142] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:16:52.325 [2024-11-19 15:22:42.659185] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:16:52.325 [2024-11-19 15:22:42.659341] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:52.325 15:22:42 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:52.325 15:22:42 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:52.585 15:22:42 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:52.585 15:22:42 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:52.585 15:22:42 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:52.585 15:22:42 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:52.585 15:22:42 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:52.585 15:22:42 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:52.585 15:22:42 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:52.585 15:22:42 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:52.585 15:22:42 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:52.585 15:22:42 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:52.585 15:22:42 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:52.585 15:22:42 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:52.585 15:22:42 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:52.585 15:22:42 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:52.585 15:22:42 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:52.585 "name": "raid_bdev1", 00:16:52.585 "uuid": "ba6c8401-0a48-4625-ae5f-b6ca4e0b0059", 00:16:52.585 "strip_size_kb": 0, 00:16:52.585 "state": "online", 00:16:52.585 "raid_level": "raid1", 00:16:52.585 "superblock": true, 00:16:52.585 "num_base_bdevs": 2, 00:16:52.585 "num_base_bdevs_discovered": 2, 00:16:52.585 "num_base_bdevs_operational": 2, 00:16:52.585 "base_bdevs_list": [ 00:16:52.585 { 00:16:52.585 "name": "pt1", 00:16:52.585 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:52.585 "is_configured": true, 00:16:52.585 "data_offset": 256, 00:16:52.585 "data_size": 7936 00:16:52.585 }, 00:16:52.585 { 00:16:52.585 "name": "pt2", 00:16:52.585 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:52.585 "is_configured": true, 00:16:52.585 "data_offset": 256, 00:16:52.585 "data_size": 7936 00:16:52.585 } 00:16:52.585 ] 00:16:52.585 }' 00:16:52.585 15:22:42 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:52.585 15:22:42 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:52.845 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:16:52.845 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:16:52.845 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:16:52.845 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:16:52.845 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@184 -- # local name 00:16:52.845 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:16:52.845 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:16:52.845 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:52.845 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:52.845 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:52.845 [2024-11-19 15:22:43.084183] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:52.846 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:52.846 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:16:52.846 "name": "raid_bdev1", 00:16:52.846 "aliases": [ 00:16:52.846 "ba6c8401-0a48-4625-ae5f-b6ca4e0b0059" 00:16:52.846 ], 00:16:52.846 "product_name": "Raid Volume", 00:16:52.846 "block_size": 4096, 00:16:52.846 "num_blocks": 7936, 00:16:52.846 "uuid": "ba6c8401-0a48-4625-ae5f-b6ca4e0b0059", 00:16:52.846 "assigned_rate_limits": { 00:16:52.846 "rw_ios_per_sec": 0, 00:16:52.846 "rw_mbytes_per_sec": 0, 00:16:52.846 "r_mbytes_per_sec": 0, 00:16:52.846 "w_mbytes_per_sec": 0 00:16:52.846 }, 00:16:52.846 "claimed": false, 00:16:52.846 "zoned": false, 00:16:52.846 "supported_io_types": { 00:16:52.846 "read": true, 00:16:52.846 "write": true, 00:16:52.846 "unmap": false, 00:16:52.846 "flush": false, 00:16:52.846 "reset": true, 00:16:52.846 "nvme_admin": false, 00:16:52.846 "nvme_io": false, 00:16:52.846 "nvme_io_md": false, 00:16:52.846 "write_zeroes": true, 00:16:52.846 "zcopy": false, 00:16:52.846 "get_zone_info": false, 00:16:52.846 "zone_management": false, 00:16:52.846 "zone_append": false, 00:16:52.846 "compare": false, 00:16:52.846 "compare_and_write": false, 00:16:52.846 "abort": false, 00:16:52.846 "seek_hole": false, 00:16:52.846 "seek_data": false, 00:16:52.846 "copy": false, 00:16:52.846 "nvme_iov_md": false 00:16:52.846 }, 00:16:52.846 "memory_domains": [ 00:16:52.846 { 00:16:52.846 "dma_device_id": "system", 00:16:52.846 "dma_device_type": 1 00:16:52.846 }, 00:16:52.846 { 00:16:52.846 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:52.846 "dma_device_type": 2 00:16:52.846 }, 00:16:52.846 { 00:16:52.846 "dma_device_id": "system", 00:16:52.846 "dma_device_type": 1 00:16:52.846 }, 00:16:52.846 { 00:16:52.846 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:52.846 "dma_device_type": 2 00:16:52.846 } 00:16:52.846 ], 00:16:52.846 "driver_specific": { 00:16:52.846 "raid": { 00:16:52.846 "uuid": "ba6c8401-0a48-4625-ae5f-b6ca4e0b0059", 00:16:52.846 "strip_size_kb": 0, 00:16:52.846 "state": "online", 00:16:52.846 "raid_level": "raid1", 00:16:52.846 "superblock": true, 00:16:52.846 "num_base_bdevs": 2, 00:16:52.846 "num_base_bdevs_discovered": 2, 00:16:52.846 "num_base_bdevs_operational": 2, 00:16:52.846 "base_bdevs_list": [ 00:16:52.846 { 00:16:52.846 "name": "pt1", 00:16:52.846 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:52.846 "is_configured": true, 00:16:52.846 "data_offset": 256, 00:16:52.846 "data_size": 7936 00:16:52.846 }, 00:16:52.846 { 00:16:52.846 "name": "pt2", 00:16:52.846 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:52.846 "is_configured": true, 00:16:52.846 "data_offset": 256, 00:16:52.846 "data_size": 7936 00:16:52.846 } 00:16:52.846 ] 00:16:52.846 } 00:16:52.846 } 00:16:52.846 }' 00:16:52.846 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:52.846 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:16:52.846 pt2' 00:16:52.846 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 ' 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:53.107 [2024-11-19 15:22:43.319682] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=ba6c8401-0a48-4625-ae5f-b6ca4e0b0059 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@436 -- # '[' -z ba6c8401-0a48-4625-ae5f-b6ca4e0b0059 ']' 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:53.107 [2024-11-19 15:22:43.367400] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:53.107 [2024-11-19 15:22:43.367427] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:53.107 [2024-11-19 15:22:43.367497] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:53.107 [2024-11-19 15:22:43.367560] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:53.107 [2024-11-19 15:22:43.367568] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:53.107 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@652 -- # local es=0 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:53.368 [2024-11-19 15:22:43.507159] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:16:53.368 [2024-11-19 15:22:43.508918] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:16:53.368 [2024-11-19 15:22:43.508984] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:16:53.368 [2024-11-19 15:22:43.509031] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:16:53.368 [2024-11-19 15:22:43.509046] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:53.368 [2024-11-19 15:22:43.509054] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:16:53.368 request: 00:16:53.368 { 00:16:53.368 "name": "raid_bdev1", 00:16:53.368 "raid_level": "raid1", 00:16:53.368 "base_bdevs": [ 00:16:53.368 "malloc1", 00:16:53.368 "malloc2" 00:16:53.368 ], 00:16:53.368 "superblock": false, 00:16:53.368 "method": "bdev_raid_create", 00:16:53.368 "req_id": 1 00:16:53.368 } 00:16:53.368 Got JSON-RPC error response 00:16:53.368 response: 00:16:53.368 { 00:16:53.368 "code": -17, 00:16:53.368 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:16:53.368 } 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@655 -- # es=1 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:53.368 [2024-11-19 15:22:43.571070] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:53.368 [2024-11-19 15:22:43.571170] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:53.368 [2024-11-19 15:22:43.571201] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:16:53.368 [2024-11-19 15:22:43.571229] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:53.368 [2024-11-19 15:22:43.573313] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:53.368 [2024-11-19 15:22:43.573396] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:53.368 [2024-11-19 15:22:43.573475] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:16:53.368 [2024-11-19 15:22:43.573542] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:53.368 pt1 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:53.368 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:53.369 "name": "raid_bdev1", 00:16:53.369 "uuid": "ba6c8401-0a48-4625-ae5f-b6ca4e0b0059", 00:16:53.369 "strip_size_kb": 0, 00:16:53.369 "state": "configuring", 00:16:53.369 "raid_level": "raid1", 00:16:53.369 "superblock": true, 00:16:53.369 "num_base_bdevs": 2, 00:16:53.369 "num_base_bdevs_discovered": 1, 00:16:53.369 "num_base_bdevs_operational": 2, 00:16:53.369 "base_bdevs_list": [ 00:16:53.369 { 00:16:53.369 "name": "pt1", 00:16:53.369 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:53.369 "is_configured": true, 00:16:53.369 "data_offset": 256, 00:16:53.369 "data_size": 7936 00:16:53.369 }, 00:16:53.369 { 00:16:53.369 "name": null, 00:16:53.369 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:53.369 "is_configured": false, 00:16:53.369 "data_offset": 256, 00:16:53.369 "data_size": 7936 00:16:53.369 } 00:16:53.369 ] 00:16:53.369 }' 00:16:53.369 15:22:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:53.369 15:22:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:53.938 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:16:53.938 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:16:53.938 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:16:53.938 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:53.938 15:22:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:53.938 15:22:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:53.938 [2024-11-19 15:22:44.054216] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:53.938 [2024-11-19 15:22:44.054259] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:53.938 [2024-11-19 15:22:44.054276] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:16:53.938 [2024-11-19 15:22:44.054284] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:53.938 [2024-11-19 15:22:44.054582] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:53.938 [2024-11-19 15:22:44.054597] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:53.938 [2024-11-19 15:22:44.054648] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:16:53.938 [2024-11-19 15:22:44.054664] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:53.938 [2024-11-19 15:22:44.054735] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:16:53.938 [2024-11-19 15:22:44.054742] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:53.938 [2024-11-19 15:22:44.054957] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:16:53.938 [2024-11-19 15:22:44.055083] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:16:53.938 [2024-11-19 15:22:44.055097] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:16:53.938 [2024-11-19 15:22:44.055184] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:53.938 pt2 00:16:53.938 15:22:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:53.938 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:16:53.939 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:16:53.939 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:53.939 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:53.939 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:53.939 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:53.939 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:53.939 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:53.939 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:53.939 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:53.939 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:53.939 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:53.939 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:53.939 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:53.939 15:22:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:53.939 15:22:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:53.939 15:22:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:53.939 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:53.939 "name": "raid_bdev1", 00:16:53.939 "uuid": "ba6c8401-0a48-4625-ae5f-b6ca4e0b0059", 00:16:53.939 "strip_size_kb": 0, 00:16:53.939 "state": "online", 00:16:53.939 "raid_level": "raid1", 00:16:53.939 "superblock": true, 00:16:53.939 "num_base_bdevs": 2, 00:16:53.939 "num_base_bdevs_discovered": 2, 00:16:53.939 "num_base_bdevs_operational": 2, 00:16:53.939 "base_bdevs_list": [ 00:16:53.939 { 00:16:53.939 "name": "pt1", 00:16:53.939 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:53.939 "is_configured": true, 00:16:53.939 "data_offset": 256, 00:16:53.939 "data_size": 7936 00:16:53.939 }, 00:16:53.939 { 00:16:53.939 "name": "pt2", 00:16:53.939 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:53.939 "is_configured": true, 00:16:53.939 "data_offset": 256, 00:16:53.939 "data_size": 7936 00:16:53.939 } 00:16:53.939 ] 00:16:53.939 }' 00:16:53.939 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:53.939 15:22:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:54.199 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:16:54.199 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:16:54.199 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:16:54.199 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:16:54.199 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@184 -- # local name 00:16:54.199 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:16:54.199 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:54.199 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:16:54.199 15:22:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:54.199 15:22:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:54.199 [2024-11-19 15:22:44.513659] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:54.199 15:22:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:54.459 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:16:54.460 "name": "raid_bdev1", 00:16:54.460 "aliases": [ 00:16:54.460 "ba6c8401-0a48-4625-ae5f-b6ca4e0b0059" 00:16:54.460 ], 00:16:54.460 "product_name": "Raid Volume", 00:16:54.460 "block_size": 4096, 00:16:54.460 "num_blocks": 7936, 00:16:54.460 "uuid": "ba6c8401-0a48-4625-ae5f-b6ca4e0b0059", 00:16:54.460 "assigned_rate_limits": { 00:16:54.460 "rw_ios_per_sec": 0, 00:16:54.460 "rw_mbytes_per_sec": 0, 00:16:54.460 "r_mbytes_per_sec": 0, 00:16:54.460 "w_mbytes_per_sec": 0 00:16:54.460 }, 00:16:54.460 "claimed": false, 00:16:54.460 "zoned": false, 00:16:54.460 "supported_io_types": { 00:16:54.460 "read": true, 00:16:54.460 "write": true, 00:16:54.460 "unmap": false, 00:16:54.460 "flush": false, 00:16:54.460 "reset": true, 00:16:54.460 "nvme_admin": false, 00:16:54.460 "nvme_io": false, 00:16:54.460 "nvme_io_md": false, 00:16:54.460 "write_zeroes": true, 00:16:54.460 "zcopy": false, 00:16:54.460 "get_zone_info": false, 00:16:54.460 "zone_management": false, 00:16:54.460 "zone_append": false, 00:16:54.460 "compare": false, 00:16:54.460 "compare_and_write": false, 00:16:54.460 "abort": false, 00:16:54.460 "seek_hole": false, 00:16:54.460 "seek_data": false, 00:16:54.460 "copy": false, 00:16:54.460 "nvme_iov_md": false 00:16:54.460 }, 00:16:54.460 "memory_domains": [ 00:16:54.460 { 00:16:54.460 "dma_device_id": "system", 00:16:54.460 "dma_device_type": 1 00:16:54.460 }, 00:16:54.460 { 00:16:54.460 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:54.460 "dma_device_type": 2 00:16:54.460 }, 00:16:54.460 { 00:16:54.460 "dma_device_id": "system", 00:16:54.460 "dma_device_type": 1 00:16:54.460 }, 00:16:54.460 { 00:16:54.460 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:54.460 "dma_device_type": 2 00:16:54.460 } 00:16:54.460 ], 00:16:54.460 "driver_specific": { 00:16:54.460 "raid": { 00:16:54.460 "uuid": "ba6c8401-0a48-4625-ae5f-b6ca4e0b0059", 00:16:54.460 "strip_size_kb": 0, 00:16:54.460 "state": "online", 00:16:54.460 "raid_level": "raid1", 00:16:54.460 "superblock": true, 00:16:54.460 "num_base_bdevs": 2, 00:16:54.460 "num_base_bdevs_discovered": 2, 00:16:54.460 "num_base_bdevs_operational": 2, 00:16:54.460 "base_bdevs_list": [ 00:16:54.460 { 00:16:54.460 "name": "pt1", 00:16:54.460 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:54.460 "is_configured": true, 00:16:54.460 "data_offset": 256, 00:16:54.460 "data_size": 7936 00:16:54.460 }, 00:16:54.460 { 00:16:54.460 "name": "pt2", 00:16:54.460 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:54.460 "is_configured": true, 00:16:54.460 "data_offset": 256, 00:16:54.460 "data_size": 7936 00:16:54.460 } 00:16:54.460 ] 00:16:54.460 } 00:16:54.460 } 00:16:54.460 }' 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:16:54.460 pt2' 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 ' 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:54.460 [2024-11-19 15:22:44.729285] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@487 -- # '[' ba6c8401-0a48-4625-ae5f-b6ca4e0b0059 '!=' ba6c8401-0a48-4625-ae5f-b6ca4e0b0059 ']' 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@198 -- # case $1 in 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@199 -- # return 0 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:54.460 [2024-11-19 15:22:44.773036] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:54.460 15:22:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:54.720 15:22:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:54.720 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:54.720 "name": "raid_bdev1", 00:16:54.720 "uuid": "ba6c8401-0a48-4625-ae5f-b6ca4e0b0059", 00:16:54.720 "strip_size_kb": 0, 00:16:54.720 "state": "online", 00:16:54.720 "raid_level": "raid1", 00:16:54.720 "superblock": true, 00:16:54.720 "num_base_bdevs": 2, 00:16:54.720 "num_base_bdevs_discovered": 1, 00:16:54.720 "num_base_bdevs_operational": 1, 00:16:54.720 "base_bdevs_list": [ 00:16:54.720 { 00:16:54.720 "name": null, 00:16:54.720 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:54.720 "is_configured": false, 00:16:54.720 "data_offset": 0, 00:16:54.720 "data_size": 7936 00:16:54.720 }, 00:16:54.720 { 00:16:54.720 "name": "pt2", 00:16:54.720 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:54.720 "is_configured": true, 00:16:54.720 "data_offset": 256, 00:16:54.720 "data_size": 7936 00:16:54.720 } 00:16:54.720 ] 00:16:54.720 }' 00:16:54.720 15:22:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:54.720 15:22:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:54.981 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:54.981 15:22:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:54.981 15:22:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:54.982 [2024-11-19 15:22:45.168307] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:54.982 [2024-11-19 15:22:45.168375] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:54.982 [2024-11-19 15:22:45.168457] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:54.982 [2024-11-19 15:22:45.168498] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:54.982 [2024-11-19 15:22:45.168506] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:16:54.982 15:22:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:54.982 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:16:54.982 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:54.982 15:22:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:54.982 15:22:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:54.982 15:22:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:54.982 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:16:54.982 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:16:54.982 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:16:54.982 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:16:54.982 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:16:54.982 15:22:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:54.982 15:22:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:54.982 15:22:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:54.982 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:16:54.982 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:16:54.982 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:16:54.982 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:16:54.982 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@519 -- # i=1 00:16:54.982 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:54.982 15:22:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:54.982 15:22:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:54.982 [2024-11-19 15:22:45.224207] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:54.982 [2024-11-19 15:22:45.224255] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:54.982 [2024-11-19 15:22:45.224273] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:16:54.982 [2024-11-19 15:22:45.224280] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:54.982 [2024-11-19 15:22:45.226352] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:54.982 [2024-11-19 15:22:45.226387] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:54.982 [2024-11-19 15:22:45.226447] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:16:54.982 [2024-11-19 15:22:45.226472] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:54.982 [2024-11-19 15:22:45.226540] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:16:54.982 [2024-11-19 15:22:45.226548] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:54.982 [2024-11-19 15:22:45.226773] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:16:54.982 [2024-11-19 15:22:45.226868] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:16:54.982 [2024-11-19 15:22:45.226885] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:16:54.982 [2024-11-19 15:22:45.227012] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:54.982 pt2 00:16:54.982 15:22:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:54.982 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:54.982 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:54.982 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:54.982 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:54.982 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:54.982 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:54.982 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:54.982 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:54.982 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:54.982 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:54.982 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:54.982 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:54.982 15:22:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:54.982 15:22:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:54.982 15:22:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:54.982 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:54.982 "name": "raid_bdev1", 00:16:54.982 "uuid": "ba6c8401-0a48-4625-ae5f-b6ca4e0b0059", 00:16:54.982 "strip_size_kb": 0, 00:16:54.982 "state": "online", 00:16:54.982 "raid_level": "raid1", 00:16:54.982 "superblock": true, 00:16:54.982 "num_base_bdevs": 2, 00:16:54.982 "num_base_bdevs_discovered": 1, 00:16:54.982 "num_base_bdevs_operational": 1, 00:16:54.982 "base_bdevs_list": [ 00:16:54.982 { 00:16:54.982 "name": null, 00:16:54.982 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:54.982 "is_configured": false, 00:16:54.982 "data_offset": 256, 00:16:54.982 "data_size": 7936 00:16:54.982 }, 00:16:54.982 { 00:16:54.982 "name": "pt2", 00:16:54.982 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:54.982 "is_configured": true, 00:16:54.982 "data_offset": 256, 00:16:54.982 "data_size": 7936 00:16:54.982 } 00:16:54.982 ] 00:16:54.982 }' 00:16:54.982 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:54.982 15:22:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:55.554 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:55.554 15:22:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:55.554 15:22:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:55.554 [2024-11-19 15:22:45.675564] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:55.554 [2024-11-19 15:22:45.675631] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:55.554 [2024-11-19 15:22:45.675725] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:55.554 [2024-11-19 15:22:45.675778] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:55.554 [2024-11-19 15:22:45.675864] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:16:55.554 15:22:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:55.554 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:55.554 15:22:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:55.554 15:22:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:55.554 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:16:55.554 15:22:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:55.554 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:16:55.554 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:16:55.554 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:16:55.554 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:55.554 15:22:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:55.554 15:22:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:55.554 [2024-11-19 15:22:45.735454] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:55.554 [2024-11-19 15:22:45.735565] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:55.554 [2024-11-19 15:22:45.735595] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008d80 00:16:55.554 [2024-11-19 15:22:45.735626] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:55.554 [2024-11-19 15:22:45.737739] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:55.554 [2024-11-19 15:22:45.737826] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:55.554 [2024-11-19 15:22:45.737909] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:16:55.554 [2024-11-19 15:22:45.737978] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:55.554 [2024-11-19 15:22:45.738115] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:16:55.554 [2024-11-19 15:22:45.738184] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:55.554 [2024-11-19 15:22:45.738225] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:16:55.554 [2024-11-19 15:22:45.738303] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:55.554 [2024-11-19 15:22:45.738409] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:16:55.554 [2024-11-19 15:22:45.738452] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:55.554 [2024-11-19 15:22:45.738686] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:16:55.554 [2024-11-19 15:22:45.738834] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:16:55.554 [2024-11-19 15:22:45.738875] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:16:55.554 [2024-11-19 15:22:45.739023] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:55.554 pt1 00:16:55.554 15:22:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:55.554 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:16:55.554 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:55.554 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:55.554 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:55.554 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:55.554 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:55.554 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:55.554 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:55.554 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:55.554 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:55.554 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:55.554 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:55.554 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:55.554 15:22:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:55.554 15:22:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:55.554 15:22:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:55.554 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:55.554 "name": "raid_bdev1", 00:16:55.554 "uuid": "ba6c8401-0a48-4625-ae5f-b6ca4e0b0059", 00:16:55.554 "strip_size_kb": 0, 00:16:55.554 "state": "online", 00:16:55.554 "raid_level": "raid1", 00:16:55.554 "superblock": true, 00:16:55.554 "num_base_bdevs": 2, 00:16:55.554 "num_base_bdevs_discovered": 1, 00:16:55.554 "num_base_bdevs_operational": 1, 00:16:55.554 "base_bdevs_list": [ 00:16:55.554 { 00:16:55.554 "name": null, 00:16:55.554 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:55.554 "is_configured": false, 00:16:55.554 "data_offset": 256, 00:16:55.554 "data_size": 7936 00:16:55.554 }, 00:16:55.554 { 00:16:55.554 "name": "pt2", 00:16:55.554 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:55.554 "is_configured": true, 00:16:55.554 "data_offset": 256, 00:16:55.554 "data_size": 7936 00:16:55.554 } 00:16:55.554 ] 00:16:55.554 }' 00:16:55.554 15:22:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:55.554 15:22:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:56.125 15:22:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:16:56.125 15:22:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:16:56.125 15:22:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:56.125 15:22:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:56.125 15:22:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:56.125 15:22:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:16:56.125 15:22:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:56.125 15:22:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:16:56.125 15:22:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:56.125 15:22:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:56.125 [2024-11-19 15:22:46.282733] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:56.125 15:22:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:56.125 15:22:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@558 -- # '[' ba6c8401-0a48-4625-ae5f-b6ca4e0b0059 '!=' ba6c8401-0a48-4625-ae5f-b6ca4e0b0059 ']' 00:16:56.125 15:22:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@563 -- # killprocess 96564 00:16:56.125 15:22:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@954 -- # '[' -z 96564 ']' 00:16:56.125 15:22:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@958 -- # kill -0 96564 00:16:56.125 15:22:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@959 -- # uname 00:16:56.125 15:22:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:16:56.125 15:22:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 96564 00:16:56.125 15:22:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:16:56.125 15:22:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:16:56.125 15:22:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@972 -- # echo 'killing process with pid 96564' 00:16:56.125 killing process with pid 96564 00:16:56.125 15:22:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@973 -- # kill 96564 00:16:56.125 [2024-11-19 15:22:46.362922] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:56.125 [2024-11-19 15:22:46.362990] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:56.125 [2024-11-19 15:22:46.363026] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:56.125 [2024-11-19 15:22:46.363033] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:16:56.125 15:22:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@978 -- # wait 96564 00:16:56.125 [2024-11-19 15:22:46.385772] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:56.385 15:22:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@565 -- # return 0 00:16:56.385 00:16:56.385 real 0m4.923s 00:16:56.385 user 0m8.024s 00:16:56.385 sys 0m1.137s 00:16:56.385 15:22:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:56.385 ************************************ 00:16:56.385 END TEST raid_superblock_test_4k 00:16:56.385 ************************************ 00:16:56.385 15:22:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:56.385 15:22:46 bdev_raid -- bdev/bdev_raid.sh@999 -- # '[' true = true ']' 00:16:56.385 15:22:46 bdev_raid -- bdev/bdev_raid.sh@1000 -- # run_test raid_rebuild_test_sb_4k raid_rebuild_test raid1 2 true false true 00:16:56.385 15:22:46 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:16:56.385 15:22:46 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:56.385 15:22:46 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:56.385 ************************************ 00:16:56.385 START TEST raid_rebuild_test_sb_4k 00:16:56.385 ************************************ 00:16:56.386 15:22:46 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 true false true 00:16:56.386 15:22:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:16:56.386 15:22:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:16:56.386 15:22:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:16:56.386 15:22:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:16:56.386 15:22:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@573 -- # local verify=true 00:16:56.386 15:22:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:16:56.386 15:22:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:56.386 15:22:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:16:56.386 15:22:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:16:56.386 15:22:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:56.386 15:22:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:16:56.386 15:22:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:16:56.386 15:22:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:56.386 15:22:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:16:56.386 15:22:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:16:56.386 15:22:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:16:56.386 15:22:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@576 -- # local strip_size 00:16:56.386 15:22:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@577 -- # local create_arg 00:16:56.386 15:22:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:16:56.386 15:22:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@579 -- # local data_offset 00:16:56.386 15:22:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:16:56.386 15:22:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:16:56.386 15:22:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:16:56.386 15:22:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:16:56.386 15:22:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@597 -- # raid_pid=96874 00:16:56.386 15:22:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:16:56.386 15:22:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@598 -- # waitforlisten 96874 00:16:56.386 15:22:46 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@835 -- # '[' -z 96874 ']' 00:16:56.386 15:22:46 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:56.386 15:22:46 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@840 -- # local max_retries=100 00:16:56.386 15:22:46 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:56.386 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:56.386 15:22:46 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@844 -- # xtrace_disable 00:16:56.386 15:22:46 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:56.646 [2024-11-19 15:22:46.774086] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:16:56.646 [2024-11-19 15:22:46.774266] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.ealI/O size of 3145728 is greater than zero copy threshold (65536). 00:16:56.646 Zero copy mechanism will not be used. 00:16:56.646 :6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid96874 ] 00:16:56.646 [2024-11-19 15:22:46.930997] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:56.646 [2024-11-19 15:22:46.956367] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:56.906 [2024-11-19 15:22:46.999517] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:56.906 [2024-11-19 15:22:46.999548] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:57.476 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:16:57.476 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@868 -- # return 0 00:16:57.476 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:16:57.476 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev1_malloc 00:16:57.476 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:57.476 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:57.476 BaseBdev1_malloc 00:16:57.476 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:57.476 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:16:57.476 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:57.476 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:57.476 [2024-11-19 15:22:47.597804] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:16:57.476 [2024-11-19 15:22:47.597902] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:57.476 [2024-11-19 15:22:47.597934] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:16:57.476 [2024-11-19 15:22:47.597953] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:57.476 [2024-11-19 15:22:47.600108] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:57.476 [2024-11-19 15:22:47.600147] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:16:57.476 BaseBdev1 00:16:57.476 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:57.476 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:16:57.476 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev2_malloc 00:16:57.476 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:57.476 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:57.476 BaseBdev2_malloc 00:16:57.476 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:57.476 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:16:57.476 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:57.476 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:57.476 [2024-11-19 15:22:47.626318] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:16:57.476 [2024-11-19 15:22:47.626362] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:57.476 [2024-11-19 15:22:47.626380] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:16:57.476 [2024-11-19 15:22:47.626387] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:57.476 [2024-11-19 15:22:47.628406] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:57.476 [2024-11-19 15:22:47.628526] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:16:57.476 BaseBdev2 00:16:57.476 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:57.477 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 4096 -b spare_malloc 00:16:57.477 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:57.477 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:57.477 spare_malloc 00:16:57.477 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:57.477 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:16:57.477 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:57.477 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:57.477 spare_delay 00:16:57.477 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:57.477 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:57.477 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:57.477 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:57.477 [2024-11-19 15:22:47.666774] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:57.477 [2024-11-19 15:22:47.666863] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:57.477 [2024-11-19 15:22:47.666886] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:16:57.477 [2024-11-19 15:22:47.666895] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:57.477 [2024-11-19 15:22:47.669007] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:57.477 [2024-11-19 15:22:47.669037] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:57.477 spare 00:16:57.477 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:57.477 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:16:57.477 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:57.477 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:57.477 [2024-11-19 15:22:47.678815] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:57.477 [2024-11-19 15:22:47.680662] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:57.477 [2024-11-19 15:22:47.680875] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:16:57.477 [2024-11-19 15:22:47.680920] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:57.477 [2024-11-19 15:22:47.681190] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:16:57.477 [2024-11-19 15:22:47.681368] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:16:57.477 [2024-11-19 15:22:47.681414] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:16:57.477 [2024-11-19 15:22:47.681570] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:57.477 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:57.477 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:57.477 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:57.477 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:57.477 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:57.477 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:57.477 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:57.477 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:57.477 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:57.477 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:57.477 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:57.477 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:57.477 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:57.477 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:57.477 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:57.477 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:57.477 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:57.477 "name": "raid_bdev1", 00:16:57.477 "uuid": "a49c1027-9619-4136-b2c0-3709f253da7b", 00:16:57.477 "strip_size_kb": 0, 00:16:57.477 "state": "online", 00:16:57.477 "raid_level": "raid1", 00:16:57.477 "superblock": true, 00:16:57.477 "num_base_bdevs": 2, 00:16:57.477 "num_base_bdevs_discovered": 2, 00:16:57.477 "num_base_bdevs_operational": 2, 00:16:57.477 "base_bdevs_list": [ 00:16:57.477 { 00:16:57.477 "name": "BaseBdev1", 00:16:57.477 "uuid": "0ca01472-b9e3-5ed0-acf0-6ad13fae8b8e", 00:16:57.477 "is_configured": true, 00:16:57.477 "data_offset": 256, 00:16:57.477 "data_size": 7936 00:16:57.477 }, 00:16:57.477 { 00:16:57.477 "name": "BaseBdev2", 00:16:57.477 "uuid": "946223a0-8046-550a-94d0-a2cfe3831db5", 00:16:57.477 "is_configured": true, 00:16:57.477 "data_offset": 256, 00:16:57.477 "data_size": 7936 00:16:57.477 } 00:16:57.477 ] 00:16:57.477 }' 00:16:57.477 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:57.477 15:22:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:58.047 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:58.047 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:16:58.047 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:58.047 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:58.047 [2024-11-19 15:22:48.166160] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:58.047 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:58.047 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=7936 00:16:58.047 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:58.047 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:16:58.047 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:58.047 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:58.047 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:58.047 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@619 -- # data_offset=256 00:16:58.047 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:16:58.047 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:16:58.047 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:16:58.047 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:16:58.047 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:16:58.047 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:16:58.047 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # local bdev_list 00:16:58.047 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:16:58.047 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # local nbd_list 00:16:58.047 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@12 -- # local i 00:16:58.047 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:16:58.047 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:16:58.048 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:16:58.307 [2024-11-19 15:22:48.433528] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:16:58.308 /dev/nbd0 00:16:58.308 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:16:58.308 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:16:58.308 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:16:58.308 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@873 -- # local i 00:16:58.308 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:16:58.308 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:16:58.308 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:16:58.308 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@877 -- # break 00:16:58.308 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:16:58.308 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:16:58.308 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:58.308 1+0 records in 00:16:58.308 1+0 records out 00:16:58.308 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000478914 s, 8.6 MB/s 00:16:58.308 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:58.308 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # size=4096 00:16:58.308 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:58.308 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:16:58.308 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@893 -- # return 0 00:16:58.308 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:16:58.308 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:16:58.308 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:16:58.308 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:16:58.308 15:22:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=4096 count=7936 oflag=direct 00:16:58.877 7936+0 records in 00:16:58.877 7936+0 records out 00:16:58.877 32505856 bytes (33 MB, 31 MiB) copied, 0.592561 s, 54.9 MB/s 00:16:58.877 15:22:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:16:58.877 15:22:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:16:58.877 15:22:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:16:58.877 15:22:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # local nbd_list 00:16:58.877 15:22:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@51 -- # local i 00:16:58.877 15:22:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:58.877 15:22:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:16:59.137 15:22:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:16:59.137 15:22:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:16:59.137 15:22:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:16:59.137 [2024-11-19 15:22:49.310932] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:59.137 15:22:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:59.137 15:22:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:59.137 15:22:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:16:59.137 15:22:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@41 -- # break 00:16:59.137 15:22:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@45 -- # return 0 00:16:59.137 15:22:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:16:59.137 15:22:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:59.137 15:22:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:59.137 [2024-11-19 15:22:49.323016] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:16:59.137 15:22:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:59.137 15:22:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:59.137 15:22:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:59.137 15:22:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:59.137 15:22:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:59.137 15:22:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:59.137 15:22:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:59.137 15:22:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:59.137 15:22:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:59.137 15:22:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:59.137 15:22:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:59.137 15:22:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:59.137 15:22:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:59.137 15:22:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:59.137 15:22:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:59.137 15:22:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:59.137 15:22:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:59.137 "name": "raid_bdev1", 00:16:59.137 "uuid": "a49c1027-9619-4136-b2c0-3709f253da7b", 00:16:59.137 "strip_size_kb": 0, 00:16:59.137 "state": "online", 00:16:59.137 "raid_level": "raid1", 00:16:59.137 "superblock": true, 00:16:59.137 "num_base_bdevs": 2, 00:16:59.137 "num_base_bdevs_discovered": 1, 00:16:59.137 "num_base_bdevs_operational": 1, 00:16:59.137 "base_bdevs_list": [ 00:16:59.137 { 00:16:59.137 "name": null, 00:16:59.137 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:59.137 "is_configured": false, 00:16:59.137 "data_offset": 0, 00:16:59.137 "data_size": 7936 00:16:59.137 }, 00:16:59.137 { 00:16:59.137 "name": "BaseBdev2", 00:16:59.137 "uuid": "946223a0-8046-550a-94d0-a2cfe3831db5", 00:16:59.137 "is_configured": true, 00:16:59.137 "data_offset": 256, 00:16:59.137 "data_size": 7936 00:16:59.137 } 00:16:59.137 ] 00:16:59.137 }' 00:16:59.137 15:22:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:59.137 15:22:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:59.707 15:22:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:59.707 15:22:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:59.707 15:22:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:59.707 [2024-11-19 15:22:49.826118] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:59.707 [2024-11-19 15:22:49.840108] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00019c960 00:16:59.707 15:22:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:59.707 15:22:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@647 -- # sleep 1 00:16:59.707 [2024-11-19 15:22:49.842256] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:17:00.646 15:22:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:00.646 15:22:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:00.646 15:22:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:00.646 15:22:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:00.646 15:22:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:00.646 15:22:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:00.647 15:22:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:00.647 15:22:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:00.647 15:22:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:00.647 15:22:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:00.647 15:22:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:00.647 "name": "raid_bdev1", 00:17:00.647 "uuid": "a49c1027-9619-4136-b2c0-3709f253da7b", 00:17:00.647 "strip_size_kb": 0, 00:17:00.647 "state": "online", 00:17:00.647 "raid_level": "raid1", 00:17:00.647 "superblock": true, 00:17:00.647 "num_base_bdevs": 2, 00:17:00.647 "num_base_bdevs_discovered": 2, 00:17:00.647 "num_base_bdevs_operational": 2, 00:17:00.647 "process": { 00:17:00.647 "type": "rebuild", 00:17:00.647 "target": "spare", 00:17:00.647 "progress": { 00:17:00.647 "blocks": 2560, 00:17:00.647 "percent": 32 00:17:00.647 } 00:17:00.647 }, 00:17:00.647 "base_bdevs_list": [ 00:17:00.647 { 00:17:00.647 "name": "spare", 00:17:00.647 "uuid": "63d9c813-6fcb-5bcf-9435-8d46d0319836", 00:17:00.647 "is_configured": true, 00:17:00.647 "data_offset": 256, 00:17:00.647 "data_size": 7936 00:17:00.647 }, 00:17:00.647 { 00:17:00.647 "name": "BaseBdev2", 00:17:00.647 "uuid": "946223a0-8046-550a-94d0-a2cfe3831db5", 00:17:00.647 "is_configured": true, 00:17:00.647 "data_offset": 256, 00:17:00.647 "data_size": 7936 00:17:00.647 } 00:17:00.647 ] 00:17:00.647 }' 00:17:00.647 15:22:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:00.647 15:22:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:00.647 15:22:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:00.907 15:22:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:00.907 15:22:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:17:00.907 15:22:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:00.907 15:22:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:00.907 [2024-11-19 15:22:50.989800] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:00.907 [2024-11-19 15:22:51.046668] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:17:00.907 [2024-11-19 15:22:51.046720] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:00.907 [2024-11-19 15:22:51.046737] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:00.907 [2024-11-19 15:22:51.046744] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:17:00.907 15:22:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:00.907 15:22:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:00.907 15:22:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:00.907 15:22:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:00.907 15:22:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:00.907 15:22:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:00.907 15:22:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:00.907 15:22:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:00.907 15:22:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:00.907 15:22:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:00.907 15:22:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:00.907 15:22:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:00.907 15:22:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:00.907 15:22:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:00.907 15:22:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:00.907 15:22:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:00.907 15:22:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:00.907 "name": "raid_bdev1", 00:17:00.907 "uuid": "a49c1027-9619-4136-b2c0-3709f253da7b", 00:17:00.907 "strip_size_kb": 0, 00:17:00.907 "state": "online", 00:17:00.907 "raid_level": "raid1", 00:17:00.907 "superblock": true, 00:17:00.907 "num_base_bdevs": 2, 00:17:00.907 "num_base_bdevs_discovered": 1, 00:17:00.907 "num_base_bdevs_operational": 1, 00:17:00.907 "base_bdevs_list": [ 00:17:00.907 { 00:17:00.907 "name": null, 00:17:00.907 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:00.907 "is_configured": false, 00:17:00.907 "data_offset": 0, 00:17:00.907 "data_size": 7936 00:17:00.907 }, 00:17:00.907 { 00:17:00.907 "name": "BaseBdev2", 00:17:00.907 "uuid": "946223a0-8046-550a-94d0-a2cfe3831db5", 00:17:00.907 "is_configured": true, 00:17:00.907 "data_offset": 256, 00:17:00.907 "data_size": 7936 00:17:00.907 } 00:17:00.907 ] 00:17:00.907 }' 00:17:00.907 15:22:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:00.907 15:22:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:01.166 15:22:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:01.166 15:22:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:01.166 15:22:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:01.166 15:22:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:01.166 15:22:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:01.166 15:22:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:01.166 15:22:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:01.166 15:22:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:01.166 15:22:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:01.426 15:22:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:01.426 15:22:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:01.426 "name": "raid_bdev1", 00:17:01.427 "uuid": "a49c1027-9619-4136-b2c0-3709f253da7b", 00:17:01.427 "strip_size_kb": 0, 00:17:01.427 "state": "online", 00:17:01.427 "raid_level": "raid1", 00:17:01.427 "superblock": true, 00:17:01.427 "num_base_bdevs": 2, 00:17:01.427 "num_base_bdevs_discovered": 1, 00:17:01.427 "num_base_bdevs_operational": 1, 00:17:01.427 "base_bdevs_list": [ 00:17:01.427 { 00:17:01.427 "name": null, 00:17:01.427 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:01.427 "is_configured": false, 00:17:01.427 "data_offset": 0, 00:17:01.427 "data_size": 7936 00:17:01.427 }, 00:17:01.427 { 00:17:01.427 "name": "BaseBdev2", 00:17:01.427 "uuid": "946223a0-8046-550a-94d0-a2cfe3831db5", 00:17:01.427 "is_configured": true, 00:17:01.427 "data_offset": 256, 00:17:01.427 "data_size": 7936 00:17:01.427 } 00:17:01.427 ] 00:17:01.427 }' 00:17:01.427 15:22:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:01.427 15:22:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:01.427 15:22:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:01.427 15:22:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:01.427 15:22:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:17:01.427 15:22:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:01.427 15:22:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:01.427 [2024-11-19 15:22:51.638282] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:17:01.427 [2024-11-19 15:22:51.642916] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00019ca30 00:17:01.427 15:22:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:01.427 15:22:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@663 -- # sleep 1 00:17:01.427 [2024-11-19 15:22:51.644813] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:17:02.366 15:22:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:02.366 15:22:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:02.366 15:22:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:02.366 15:22:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:02.366 15:22:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:02.366 15:22:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:02.366 15:22:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:02.366 15:22:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:02.366 15:22:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:02.366 15:22:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:02.366 15:22:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:02.366 "name": "raid_bdev1", 00:17:02.366 "uuid": "a49c1027-9619-4136-b2c0-3709f253da7b", 00:17:02.366 "strip_size_kb": 0, 00:17:02.366 "state": "online", 00:17:02.366 "raid_level": "raid1", 00:17:02.366 "superblock": true, 00:17:02.366 "num_base_bdevs": 2, 00:17:02.366 "num_base_bdevs_discovered": 2, 00:17:02.366 "num_base_bdevs_operational": 2, 00:17:02.366 "process": { 00:17:02.366 "type": "rebuild", 00:17:02.366 "target": "spare", 00:17:02.366 "progress": { 00:17:02.366 "blocks": 2560, 00:17:02.366 "percent": 32 00:17:02.366 } 00:17:02.366 }, 00:17:02.366 "base_bdevs_list": [ 00:17:02.366 { 00:17:02.366 "name": "spare", 00:17:02.366 "uuid": "63d9c813-6fcb-5bcf-9435-8d46d0319836", 00:17:02.366 "is_configured": true, 00:17:02.366 "data_offset": 256, 00:17:02.366 "data_size": 7936 00:17:02.366 }, 00:17:02.366 { 00:17:02.366 "name": "BaseBdev2", 00:17:02.366 "uuid": "946223a0-8046-550a-94d0-a2cfe3831db5", 00:17:02.366 "is_configured": true, 00:17:02.366 "data_offset": 256, 00:17:02.366 "data_size": 7936 00:17:02.366 } 00:17:02.366 ] 00:17:02.366 }' 00:17:02.366 15:22:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:02.641 15:22:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:02.641 15:22:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:02.641 15:22:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:02.641 15:22:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:17:02.641 15:22:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:17:02.641 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:17:02.641 15:22:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:17:02.641 15:22:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:17:02.641 15:22:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:17:02.641 15:22:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@706 -- # local timeout=570 00:17:02.641 15:22:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:17:02.641 15:22:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:02.641 15:22:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:02.641 15:22:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:02.641 15:22:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:02.641 15:22:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:02.641 15:22:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:02.641 15:22:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:02.641 15:22:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:02.641 15:22:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:02.641 15:22:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:02.641 15:22:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:02.641 "name": "raid_bdev1", 00:17:02.641 "uuid": "a49c1027-9619-4136-b2c0-3709f253da7b", 00:17:02.641 "strip_size_kb": 0, 00:17:02.641 "state": "online", 00:17:02.641 "raid_level": "raid1", 00:17:02.641 "superblock": true, 00:17:02.641 "num_base_bdevs": 2, 00:17:02.641 "num_base_bdevs_discovered": 2, 00:17:02.641 "num_base_bdevs_operational": 2, 00:17:02.641 "process": { 00:17:02.641 "type": "rebuild", 00:17:02.641 "target": "spare", 00:17:02.641 "progress": { 00:17:02.641 "blocks": 2816, 00:17:02.641 "percent": 35 00:17:02.641 } 00:17:02.641 }, 00:17:02.641 "base_bdevs_list": [ 00:17:02.641 { 00:17:02.641 "name": "spare", 00:17:02.641 "uuid": "63d9c813-6fcb-5bcf-9435-8d46d0319836", 00:17:02.641 "is_configured": true, 00:17:02.641 "data_offset": 256, 00:17:02.641 "data_size": 7936 00:17:02.641 }, 00:17:02.641 { 00:17:02.641 "name": "BaseBdev2", 00:17:02.641 "uuid": "946223a0-8046-550a-94d0-a2cfe3831db5", 00:17:02.641 "is_configured": true, 00:17:02.641 "data_offset": 256, 00:17:02.641 "data_size": 7936 00:17:02.641 } 00:17:02.641 ] 00:17:02.641 }' 00:17:02.641 15:22:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:02.641 15:22:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:02.641 15:22:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:02.641 15:22:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:02.641 15:22:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@711 -- # sleep 1 00:17:04.037 15:22:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:17:04.037 15:22:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:04.037 15:22:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:04.037 15:22:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:04.037 15:22:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:04.037 15:22:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:04.037 15:22:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:04.037 15:22:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:04.037 15:22:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:04.037 15:22:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:04.038 15:22:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:04.038 15:22:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:04.038 "name": "raid_bdev1", 00:17:04.038 "uuid": "a49c1027-9619-4136-b2c0-3709f253da7b", 00:17:04.038 "strip_size_kb": 0, 00:17:04.038 "state": "online", 00:17:04.038 "raid_level": "raid1", 00:17:04.038 "superblock": true, 00:17:04.038 "num_base_bdevs": 2, 00:17:04.038 "num_base_bdevs_discovered": 2, 00:17:04.038 "num_base_bdevs_operational": 2, 00:17:04.038 "process": { 00:17:04.038 "type": "rebuild", 00:17:04.038 "target": "spare", 00:17:04.038 "progress": { 00:17:04.038 "blocks": 5888, 00:17:04.038 "percent": 74 00:17:04.038 } 00:17:04.038 }, 00:17:04.038 "base_bdevs_list": [ 00:17:04.038 { 00:17:04.038 "name": "spare", 00:17:04.038 "uuid": "63d9c813-6fcb-5bcf-9435-8d46d0319836", 00:17:04.038 "is_configured": true, 00:17:04.038 "data_offset": 256, 00:17:04.038 "data_size": 7936 00:17:04.038 }, 00:17:04.038 { 00:17:04.038 "name": "BaseBdev2", 00:17:04.038 "uuid": "946223a0-8046-550a-94d0-a2cfe3831db5", 00:17:04.038 "is_configured": true, 00:17:04.038 "data_offset": 256, 00:17:04.038 "data_size": 7936 00:17:04.038 } 00:17:04.038 ] 00:17:04.038 }' 00:17:04.038 15:22:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:04.038 15:22:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:04.038 15:22:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:04.038 15:22:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:04.038 15:22:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@711 -- # sleep 1 00:17:04.608 [2024-11-19 15:22:54.754764] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:17:04.608 [2024-11-19 15:22:54.754840] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:17:04.608 [2024-11-19 15:22:54.754932] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:04.868 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:17:04.868 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:04.868 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:04.868 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:04.868 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:04.868 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:04.868 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:04.868 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:04.868 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:04.868 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:04.868 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:04.868 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:04.868 "name": "raid_bdev1", 00:17:04.868 "uuid": "a49c1027-9619-4136-b2c0-3709f253da7b", 00:17:04.868 "strip_size_kb": 0, 00:17:04.868 "state": "online", 00:17:04.868 "raid_level": "raid1", 00:17:04.868 "superblock": true, 00:17:04.868 "num_base_bdevs": 2, 00:17:04.868 "num_base_bdevs_discovered": 2, 00:17:04.868 "num_base_bdevs_operational": 2, 00:17:04.868 "base_bdevs_list": [ 00:17:04.868 { 00:17:04.868 "name": "spare", 00:17:04.868 "uuid": "63d9c813-6fcb-5bcf-9435-8d46d0319836", 00:17:04.868 "is_configured": true, 00:17:04.868 "data_offset": 256, 00:17:04.868 "data_size": 7936 00:17:04.868 }, 00:17:04.868 { 00:17:04.868 "name": "BaseBdev2", 00:17:04.868 "uuid": "946223a0-8046-550a-94d0-a2cfe3831db5", 00:17:04.868 "is_configured": true, 00:17:04.868 "data_offset": 256, 00:17:04.868 "data_size": 7936 00:17:04.868 } 00:17:04.868 ] 00:17:04.868 }' 00:17:04.868 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:04.868 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:17:05.129 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:05.129 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:17:05.129 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@709 -- # break 00:17:05.129 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:05.129 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:05.129 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:05.129 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:05.129 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:05.129 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:05.129 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:05.129 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:05.129 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:05.129 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:05.129 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:05.129 "name": "raid_bdev1", 00:17:05.129 "uuid": "a49c1027-9619-4136-b2c0-3709f253da7b", 00:17:05.129 "strip_size_kb": 0, 00:17:05.129 "state": "online", 00:17:05.129 "raid_level": "raid1", 00:17:05.129 "superblock": true, 00:17:05.129 "num_base_bdevs": 2, 00:17:05.129 "num_base_bdevs_discovered": 2, 00:17:05.129 "num_base_bdevs_operational": 2, 00:17:05.129 "base_bdevs_list": [ 00:17:05.129 { 00:17:05.129 "name": "spare", 00:17:05.129 "uuid": "63d9c813-6fcb-5bcf-9435-8d46d0319836", 00:17:05.129 "is_configured": true, 00:17:05.129 "data_offset": 256, 00:17:05.129 "data_size": 7936 00:17:05.129 }, 00:17:05.129 { 00:17:05.129 "name": "BaseBdev2", 00:17:05.129 "uuid": "946223a0-8046-550a-94d0-a2cfe3831db5", 00:17:05.129 "is_configured": true, 00:17:05.129 "data_offset": 256, 00:17:05.129 "data_size": 7936 00:17:05.129 } 00:17:05.129 ] 00:17:05.129 }' 00:17:05.129 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:05.129 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:05.129 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:05.129 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:05.129 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:17:05.129 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:05.129 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:05.129 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:05.129 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:05.129 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:05.129 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:05.129 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:05.129 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:05.129 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:05.129 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:05.129 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:05.129 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:05.129 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:05.129 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:05.129 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:05.129 "name": "raid_bdev1", 00:17:05.129 "uuid": "a49c1027-9619-4136-b2c0-3709f253da7b", 00:17:05.129 "strip_size_kb": 0, 00:17:05.129 "state": "online", 00:17:05.129 "raid_level": "raid1", 00:17:05.129 "superblock": true, 00:17:05.129 "num_base_bdevs": 2, 00:17:05.129 "num_base_bdevs_discovered": 2, 00:17:05.129 "num_base_bdevs_operational": 2, 00:17:05.129 "base_bdevs_list": [ 00:17:05.129 { 00:17:05.129 "name": "spare", 00:17:05.129 "uuid": "63d9c813-6fcb-5bcf-9435-8d46d0319836", 00:17:05.129 "is_configured": true, 00:17:05.129 "data_offset": 256, 00:17:05.129 "data_size": 7936 00:17:05.129 }, 00:17:05.129 { 00:17:05.129 "name": "BaseBdev2", 00:17:05.129 "uuid": "946223a0-8046-550a-94d0-a2cfe3831db5", 00:17:05.129 "is_configured": true, 00:17:05.129 "data_offset": 256, 00:17:05.129 "data_size": 7936 00:17:05.129 } 00:17:05.129 ] 00:17:05.129 }' 00:17:05.129 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:05.129 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:05.699 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:17:05.699 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:05.699 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:05.699 [2024-11-19 15:22:55.861442] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:17:05.699 [2024-11-19 15:22:55.861477] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:17:05.699 [2024-11-19 15:22:55.861553] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:05.699 [2024-11-19 15:22:55.861617] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:05.699 [2024-11-19 15:22:55.861631] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:17:05.699 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:05.699 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:05.699 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:05.699 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:05.699 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@720 -- # jq length 00:17:05.699 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:05.699 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:17:05.699 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:17:05.699 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:17:05.699 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:17:05.699 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:17:05.699 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:17:05.699 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # local bdev_list 00:17:05.699 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:17:05.699 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # local nbd_list 00:17:05.699 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@12 -- # local i 00:17:05.699 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:17:05.699 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:17:05.699 15:22:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:17:05.959 /dev/nbd0 00:17:05.959 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:17:05.959 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:17:05.959 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:17:05.959 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@873 -- # local i 00:17:05.959 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:17:05.959 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:17:05.959 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:17:05.959 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@877 -- # break 00:17:05.959 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:17:05.959 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:17:05.959 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:17:05.959 1+0 records in 00:17:05.959 1+0 records out 00:17:05.959 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000313965 s, 13.0 MB/s 00:17:05.959 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:17:05.959 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # size=4096 00:17:05.959 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:17:05.959 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:17:05.959 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@893 -- # return 0 00:17:05.959 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:17:05.959 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:17:05.959 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:17:06.219 /dev/nbd1 00:17:06.219 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:17:06.219 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:17:06.219 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:17:06.219 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@873 -- # local i 00:17:06.219 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:17:06.219 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:17:06.219 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:17:06.219 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@877 -- # break 00:17:06.219 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:17:06.219 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:17:06.219 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:17:06.219 1+0 records in 00:17:06.219 1+0 records out 00:17:06.219 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000349958 s, 11.7 MB/s 00:17:06.219 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:17:06.219 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # size=4096 00:17:06.219 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:17:06.219 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:17:06.219 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@893 -- # return 0 00:17:06.219 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:17:06.219 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:17:06.219 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:17:06.219 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:17:06.219 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:17:06.219 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:17:06.219 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # local nbd_list 00:17:06.219 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@51 -- # local i 00:17:06.220 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:17:06.220 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:17:06.479 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:17:06.479 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:17:06.480 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:17:06.480 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:17:06.480 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:17:06.480 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:17:06.480 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@41 -- # break 00:17:06.480 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@45 -- # return 0 00:17:06.480 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:17:06.480 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:17:06.740 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:17:06.740 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:17:06.740 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:17:06.740 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:17:06.740 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:17:06.740 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:17:06.740 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@41 -- # break 00:17:06.740 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@45 -- # return 0 00:17:06.740 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:17:06.740 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:17:06.740 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:06.740 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:06.740 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:06.740 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:17:06.740 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:06.740 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:06.740 [2024-11-19 15:22:56.915265] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:17:06.740 [2024-11-19 15:22:56.915318] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:06.740 [2024-11-19 15:22:56.915336] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:17:06.740 [2024-11-19 15:22:56.915349] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:06.740 [2024-11-19 15:22:56.917394] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:06.740 [2024-11-19 15:22:56.917438] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:17:06.740 [2024-11-19 15:22:56.917517] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:17:06.740 [2024-11-19 15:22:56.917554] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:17:06.740 [2024-11-19 15:22:56.917678] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:17:06.740 spare 00:17:06.740 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:06.740 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:17:06.740 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:06.740 15:22:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:06.740 [2024-11-19 15:22:57.017571] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:17:06.740 [2024-11-19 15:22:57.017642] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:17:06.740 [2024-11-19 15:22:57.017901] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001bb1b0 00:17:06.740 [2024-11-19 15:22:57.018059] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:17:06.740 [2024-11-19 15:22:57.018074] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:17:06.740 [2024-11-19 15:22:57.018212] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:06.740 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:06.740 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:17:06.740 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:06.740 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:06.740 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:06.740 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:06.740 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:06.740 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:06.740 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:06.740 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:06.740 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:06.740 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:06.740 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:06.740 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:06.740 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:06.740 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:06.740 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:06.740 "name": "raid_bdev1", 00:17:06.740 "uuid": "a49c1027-9619-4136-b2c0-3709f253da7b", 00:17:06.740 "strip_size_kb": 0, 00:17:06.740 "state": "online", 00:17:06.740 "raid_level": "raid1", 00:17:06.740 "superblock": true, 00:17:06.740 "num_base_bdevs": 2, 00:17:06.740 "num_base_bdevs_discovered": 2, 00:17:06.740 "num_base_bdevs_operational": 2, 00:17:06.740 "base_bdevs_list": [ 00:17:06.740 { 00:17:06.740 "name": "spare", 00:17:06.740 "uuid": "63d9c813-6fcb-5bcf-9435-8d46d0319836", 00:17:06.740 "is_configured": true, 00:17:06.740 "data_offset": 256, 00:17:06.740 "data_size": 7936 00:17:06.740 }, 00:17:06.740 { 00:17:06.740 "name": "BaseBdev2", 00:17:06.740 "uuid": "946223a0-8046-550a-94d0-a2cfe3831db5", 00:17:06.740 "is_configured": true, 00:17:06.740 "data_offset": 256, 00:17:06.740 "data_size": 7936 00:17:06.740 } 00:17:06.740 ] 00:17:06.740 }' 00:17:06.740 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:06.740 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:07.310 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:07.310 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:07.310 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:07.310 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:07.310 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:07.310 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:07.310 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:07.310 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:07.310 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:07.310 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:07.310 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:07.310 "name": "raid_bdev1", 00:17:07.310 "uuid": "a49c1027-9619-4136-b2c0-3709f253da7b", 00:17:07.310 "strip_size_kb": 0, 00:17:07.310 "state": "online", 00:17:07.310 "raid_level": "raid1", 00:17:07.310 "superblock": true, 00:17:07.310 "num_base_bdevs": 2, 00:17:07.310 "num_base_bdevs_discovered": 2, 00:17:07.310 "num_base_bdevs_operational": 2, 00:17:07.310 "base_bdevs_list": [ 00:17:07.310 { 00:17:07.310 "name": "spare", 00:17:07.310 "uuid": "63d9c813-6fcb-5bcf-9435-8d46d0319836", 00:17:07.310 "is_configured": true, 00:17:07.310 "data_offset": 256, 00:17:07.310 "data_size": 7936 00:17:07.310 }, 00:17:07.310 { 00:17:07.310 "name": "BaseBdev2", 00:17:07.310 "uuid": "946223a0-8046-550a-94d0-a2cfe3831db5", 00:17:07.310 "is_configured": true, 00:17:07.310 "data_offset": 256, 00:17:07.310 "data_size": 7936 00:17:07.310 } 00:17:07.310 ] 00:17:07.310 }' 00:17:07.311 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:07.311 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:07.311 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:07.311 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:07.311 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:07.311 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:17:07.311 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:07.311 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:07.571 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:07.571 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:17:07.571 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:17:07.571 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:07.571 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:07.571 [2024-11-19 15:22:57.694018] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:07.571 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:07.571 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:07.571 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:07.571 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:07.571 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:07.571 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:07.571 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:07.571 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:07.571 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:07.571 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:07.571 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:07.571 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:07.571 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:07.571 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:07.571 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:07.571 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:07.571 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:07.571 "name": "raid_bdev1", 00:17:07.571 "uuid": "a49c1027-9619-4136-b2c0-3709f253da7b", 00:17:07.571 "strip_size_kb": 0, 00:17:07.571 "state": "online", 00:17:07.571 "raid_level": "raid1", 00:17:07.571 "superblock": true, 00:17:07.571 "num_base_bdevs": 2, 00:17:07.571 "num_base_bdevs_discovered": 1, 00:17:07.571 "num_base_bdevs_operational": 1, 00:17:07.571 "base_bdevs_list": [ 00:17:07.571 { 00:17:07.571 "name": null, 00:17:07.571 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:07.571 "is_configured": false, 00:17:07.571 "data_offset": 0, 00:17:07.571 "data_size": 7936 00:17:07.571 }, 00:17:07.571 { 00:17:07.571 "name": "BaseBdev2", 00:17:07.571 "uuid": "946223a0-8046-550a-94d0-a2cfe3831db5", 00:17:07.571 "is_configured": true, 00:17:07.571 "data_offset": 256, 00:17:07.571 "data_size": 7936 00:17:07.571 } 00:17:07.571 ] 00:17:07.571 }' 00:17:07.571 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:07.571 15:22:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:07.831 15:22:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:17:07.831 15:22:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:07.831 15:22:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:07.831 [2024-11-19 15:22:58.137256] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:17:07.831 [2024-11-19 15:22:58.137463] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:17:07.831 [2024-11-19 15:22:58.137522] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:17:07.831 [2024-11-19 15:22:58.137590] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:17:07.831 [2024-11-19 15:22:58.142502] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001bb280 00:17:07.831 15:22:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:07.831 15:22:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@757 -- # sleep 1 00:17:07.831 [2024-11-19 15:22:58.144583] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:17:09.213 15:22:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:09.213 15:22:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:09.213 15:22:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:09.213 15:22:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:09.213 15:22:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:09.213 15:22:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:09.213 15:22:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:09.213 15:22:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:09.213 15:22:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:09.213 15:22:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:09.213 15:22:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:09.213 "name": "raid_bdev1", 00:17:09.213 "uuid": "a49c1027-9619-4136-b2c0-3709f253da7b", 00:17:09.213 "strip_size_kb": 0, 00:17:09.213 "state": "online", 00:17:09.213 "raid_level": "raid1", 00:17:09.213 "superblock": true, 00:17:09.213 "num_base_bdevs": 2, 00:17:09.213 "num_base_bdevs_discovered": 2, 00:17:09.213 "num_base_bdevs_operational": 2, 00:17:09.213 "process": { 00:17:09.213 "type": "rebuild", 00:17:09.213 "target": "spare", 00:17:09.213 "progress": { 00:17:09.213 "blocks": 2560, 00:17:09.213 "percent": 32 00:17:09.213 } 00:17:09.213 }, 00:17:09.213 "base_bdevs_list": [ 00:17:09.213 { 00:17:09.213 "name": "spare", 00:17:09.213 "uuid": "63d9c813-6fcb-5bcf-9435-8d46d0319836", 00:17:09.213 "is_configured": true, 00:17:09.213 "data_offset": 256, 00:17:09.213 "data_size": 7936 00:17:09.213 }, 00:17:09.213 { 00:17:09.213 "name": "BaseBdev2", 00:17:09.213 "uuid": "946223a0-8046-550a-94d0-a2cfe3831db5", 00:17:09.213 "is_configured": true, 00:17:09.213 "data_offset": 256, 00:17:09.213 "data_size": 7936 00:17:09.213 } 00:17:09.213 ] 00:17:09.213 }' 00:17:09.213 15:22:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:09.213 15:22:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:09.213 15:22:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:09.213 15:22:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:09.213 15:22:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:17:09.213 15:22:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:09.213 15:22:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:09.213 [2024-11-19 15:22:59.284627] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:09.213 [2024-11-19 15:22:59.348451] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:17:09.213 [2024-11-19 15:22:59.348503] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:09.213 [2024-11-19 15:22:59.348519] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:09.213 [2024-11-19 15:22:59.348526] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:17:09.213 15:22:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:09.213 15:22:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:09.214 15:22:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:09.214 15:22:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:09.214 15:22:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:09.214 15:22:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:09.214 15:22:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:09.214 15:22:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:09.214 15:22:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:09.214 15:22:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:09.214 15:22:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:09.214 15:22:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:09.214 15:22:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:09.214 15:22:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:09.214 15:22:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:09.214 15:22:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:09.214 15:22:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:09.214 "name": "raid_bdev1", 00:17:09.214 "uuid": "a49c1027-9619-4136-b2c0-3709f253da7b", 00:17:09.214 "strip_size_kb": 0, 00:17:09.214 "state": "online", 00:17:09.214 "raid_level": "raid1", 00:17:09.214 "superblock": true, 00:17:09.214 "num_base_bdevs": 2, 00:17:09.214 "num_base_bdevs_discovered": 1, 00:17:09.214 "num_base_bdevs_operational": 1, 00:17:09.214 "base_bdevs_list": [ 00:17:09.214 { 00:17:09.214 "name": null, 00:17:09.214 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:09.214 "is_configured": false, 00:17:09.214 "data_offset": 0, 00:17:09.214 "data_size": 7936 00:17:09.214 }, 00:17:09.214 { 00:17:09.214 "name": "BaseBdev2", 00:17:09.214 "uuid": "946223a0-8046-550a-94d0-a2cfe3831db5", 00:17:09.214 "is_configured": true, 00:17:09.214 "data_offset": 256, 00:17:09.214 "data_size": 7936 00:17:09.214 } 00:17:09.214 ] 00:17:09.214 }' 00:17:09.214 15:22:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:09.214 15:22:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:09.474 15:22:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:17:09.474 15:22:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:09.474 15:22:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:09.474 [2024-11-19 15:22:59.760231] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:17:09.474 [2024-11-19 15:22:59.760329] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:09.474 [2024-11-19 15:22:59.760367] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:17:09.474 [2024-11-19 15:22:59.760396] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:09.474 [2024-11-19 15:22:59.760831] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:09.474 [2024-11-19 15:22:59.760888] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:17:09.474 [2024-11-19 15:22:59.761004] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:17:09.474 [2024-11-19 15:22:59.761046] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:17:09.474 [2024-11-19 15:22:59.761109] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:17:09.474 [2024-11-19 15:22:59.761162] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:17:09.474 [2024-11-19 15:22:59.765034] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001bb350 00:17:09.474 spare 00:17:09.474 15:22:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:09.474 15:22:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@764 -- # sleep 1 00:17:09.474 [2024-11-19 15:22:59.766913] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:17:10.857 15:23:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:10.857 15:23:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:10.857 15:23:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:10.857 15:23:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:10.857 15:23:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:10.857 15:23:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:10.857 15:23:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:10.857 15:23:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:10.857 15:23:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:10.857 15:23:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:10.857 15:23:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:10.857 "name": "raid_bdev1", 00:17:10.857 "uuid": "a49c1027-9619-4136-b2c0-3709f253da7b", 00:17:10.857 "strip_size_kb": 0, 00:17:10.857 "state": "online", 00:17:10.857 "raid_level": "raid1", 00:17:10.857 "superblock": true, 00:17:10.857 "num_base_bdevs": 2, 00:17:10.857 "num_base_bdevs_discovered": 2, 00:17:10.857 "num_base_bdevs_operational": 2, 00:17:10.857 "process": { 00:17:10.857 "type": "rebuild", 00:17:10.857 "target": "spare", 00:17:10.857 "progress": { 00:17:10.857 "blocks": 2560, 00:17:10.857 "percent": 32 00:17:10.857 } 00:17:10.857 }, 00:17:10.857 "base_bdevs_list": [ 00:17:10.857 { 00:17:10.857 "name": "spare", 00:17:10.857 "uuid": "63d9c813-6fcb-5bcf-9435-8d46d0319836", 00:17:10.857 "is_configured": true, 00:17:10.857 "data_offset": 256, 00:17:10.857 "data_size": 7936 00:17:10.857 }, 00:17:10.857 { 00:17:10.857 "name": "BaseBdev2", 00:17:10.857 "uuid": "946223a0-8046-550a-94d0-a2cfe3831db5", 00:17:10.857 "is_configured": true, 00:17:10.857 "data_offset": 256, 00:17:10.857 "data_size": 7936 00:17:10.857 } 00:17:10.857 ] 00:17:10.857 }' 00:17:10.857 15:23:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:10.857 15:23:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:10.857 15:23:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:10.857 15:23:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:10.857 15:23:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:17:10.857 15:23:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:10.857 15:23:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:10.857 [2024-11-19 15:23:00.903045] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:10.857 [2024-11-19 15:23:00.970754] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:17:10.857 [2024-11-19 15:23:00.970868] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:10.857 [2024-11-19 15:23:00.970903] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:10.857 [2024-11-19 15:23:00.970940] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:17:10.857 15:23:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:10.857 15:23:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:10.857 15:23:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:10.857 15:23:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:10.857 15:23:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:10.857 15:23:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:10.857 15:23:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:10.857 15:23:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:10.857 15:23:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:10.857 15:23:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:10.857 15:23:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:10.857 15:23:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:10.857 15:23:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:10.857 15:23:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:10.857 15:23:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:10.857 15:23:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:10.857 15:23:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:10.857 "name": "raid_bdev1", 00:17:10.857 "uuid": "a49c1027-9619-4136-b2c0-3709f253da7b", 00:17:10.857 "strip_size_kb": 0, 00:17:10.857 "state": "online", 00:17:10.857 "raid_level": "raid1", 00:17:10.857 "superblock": true, 00:17:10.857 "num_base_bdevs": 2, 00:17:10.857 "num_base_bdevs_discovered": 1, 00:17:10.857 "num_base_bdevs_operational": 1, 00:17:10.857 "base_bdevs_list": [ 00:17:10.857 { 00:17:10.857 "name": null, 00:17:10.857 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:10.857 "is_configured": false, 00:17:10.857 "data_offset": 0, 00:17:10.857 "data_size": 7936 00:17:10.857 }, 00:17:10.857 { 00:17:10.857 "name": "BaseBdev2", 00:17:10.857 "uuid": "946223a0-8046-550a-94d0-a2cfe3831db5", 00:17:10.857 "is_configured": true, 00:17:10.857 "data_offset": 256, 00:17:10.857 "data_size": 7936 00:17:10.857 } 00:17:10.857 ] 00:17:10.857 }' 00:17:10.857 15:23:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:10.857 15:23:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:11.117 15:23:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:11.117 15:23:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:11.117 15:23:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:11.117 15:23:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:11.117 15:23:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:11.117 15:23:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:11.117 15:23:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:11.117 15:23:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:11.117 15:23:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:11.117 15:23:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:11.377 15:23:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:11.377 "name": "raid_bdev1", 00:17:11.377 "uuid": "a49c1027-9619-4136-b2c0-3709f253da7b", 00:17:11.377 "strip_size_kb": 0, 00:17:11.377 "state": "online", 00:17:11.377 "raid_level": "raid1", 00:17:11.377 "superblock": true, 00:17:11.377 "num_base_bdevs": 2, 00:17:11.377 "num_base_bdevs_discovered": 1, 00:17:11.377 "num_base_bdevs_operational": 1, 00:17:11.377 "base_bdevs_list": [ 00:17:11.377 { 00:17:11.377 "name": null, 00:17:11.377 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:11.377 "is_configured": false, 00:17:11.377 "data_offset": 0, 00:17:11.377 "data_size": 7936 00:17:11.377 }, 00:17:11.377 { 00:17:11.377 "name": "BaseBdev2", 00:17:11.377 "uuid": "946223a0-8046-550a-94d0-a2cfe3831db5", 00:17:11.377 "is_configured": true, 00:17:11.377 "data_offset": 256, 00:17:11.377 "data_size": 7936 00:17:11.377 } 00:17:11.377 ] 00:17:11.377 }' 00:17:11.377 15:23:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:11.377 15:23:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:11.377 15:23:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:11.377 15:23:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:11.377 15:23:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:17:11.377 15:23:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:11.378 15:23:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:11.378 15:23:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:11.378 15:23:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:17:11.378 15:23:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:11.378 15:23:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:11.378 [2024-11-19 15:23:01.598086] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:17:11.378 [2024-11-19 15:23:01.598139] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:11.378 [2024-11-19 15:23:01.598157] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:17:11.378 [2024-11-19 15:23:01.598167] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:11.378 [2024-11-19 15:23:01.598530] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:11.378 [2024-11-19 15:23:01.598549] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:17:11.378 [2024-11-19 15:23:01.598626] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:17:11.378 [2024-11-19 15:23:01.598643] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:17:11.378 [2024-11-19 15:23:01.598652] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:17:11.378 [2024-11-19 15:23:01.598665] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:17:11.378 BaseBdev1 00:17:11.378 15:23:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:11.378 15:23:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@775 -- # sleep 1 00:17:12.316 15:23:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:12.316 15:23:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:12.316 15:23:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:12.316 15:23:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:12.316 15:23:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:12.316 15:23:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:12.316 15:23:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:12.316 15:23:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:12.316 15:23:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:12.316 15:23:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:12.316 15:23:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:12.316 15:23:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:12.316 15:23:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:12.316 15:23:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:12.316 15:23:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:12.576 15:23:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:12.576 "name": "raid_bdev1", 00:17:12.576 "uuid": "a49c1027-9619-4136-b2c0-3709f253da7b", 00:17:12.576 "strip_size_kb": 0, 00:17:12.576 "state": "online", 00:17:12.576 "raid_level": "raid1", 00:17:12.576 "superblock": true, 00:17:12.576 "num_base_bdevs": 2, 00:17:12.576 "num_base_bdevs_discovered": 1, 00:17:12.576 "num_base_bdevs_operational": 1, 00:17:12.576 "base_bdevs_list": [ 00:17:12.576 { 00:17:12.576 "name": null, 00:17:12.576 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:12.576 "is_configured": false, 00:17:12.576 "data_offset": 0, 00:17:12.576 "data_size": 7936 00:17:12.576 }, 00:17:12.576 { 00:17:12.576 "name": "BaseBdev2", 00:17:12.576 "uuid": "946223a0-8046-550a-94d0-a2cfe3831db5", 00:17:12.576 "is_configured": true, 00:17:12.576 "data_offset": 256, 00:17:12.576 "data_size": 7936 00:17:12.576 } 00:17:12.576 ] 00:17:12.576 }' 00:17:12.576 15:23:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:12.576 15:23:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:12.836 15:23:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:12.836 15:23:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:12.836 15:23:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:12.836 15:23:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:12.836 15:23:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:12.836 15:23:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:12.836 15:23:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:12.836 15:23:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:12.836 15:23:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:12.836 15:23:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:12.836 15:23:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:12.836 "name": "raid_bdev1", 00:17:12.836 "uuid": "a49c1027-9619-4136-b2c0-3709f253da7b", 00:17:12.836 "strip_size_kb": 0, 00:17:12.836 "state": "online", 00:17:12.836 "raid_level": "raid1", 00:17:12.836 "superblock": true, 00:17:12.836 "num_base_bdevs": 2, 00:17:12.836 "num_base_bdevs_discovered": 1, 00:17:12.836 "num_base_bdevs_operational": 1, 00:17:12.836 "base_bdevs_list": [ 00:17:12.836 { 00:17:12.836 "name": null, 00:17:12.836 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:12.836 "is_configured": false, 00:17:12.836 "data_offset": 0, 00:17:12.836 "data_size": 7936 00:17:12.836 }, 00:17:12.836 { 00:17:12.836 "name": "BaseBdev2", 00:17:12.836 "uuid": "946223a0-8046-550a-94d0-a2cfe3831db5", 00:17:12.836 "is_configured": true, 00:17:12.836 "data_offset": 256, 00:17:12.836 "data_size": 7936 00:17:12.836 } 00:17:12.836 ] 00:17:12.836 }' 00:17:12.836 15:23:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:13.096 15:23:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:13.097 15:23:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:13.097 15:23:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:13.097 15:23:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:17:13.097 15:23:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@652 -- # local es=0 00:17:13.097 15:23:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:17:13.097 15:23:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:17:13.097 15:23:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:17:13.097 15:23:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:17:13.097 15:23:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:17:13.097 15:23:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:17:13.097 15:23:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:13.097 15:23:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:13.097 [2024-11-19 15:23:03.239452] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:17:13.097 [2024-11-19 15:23:03.239622] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:17:13.097 [2024-11-19 15:23:03.239691] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:17:13.097 request: 00:17:13.097 { 00:17:13.097 "base_bdev": "BaseBdev1", 00:17:13.097 "raid_bdev": "raid_bdev1", 00:17:13.097 "method": "bdev_raid_add_base_bdev", 00:17:13.097 "req_id": 1 00:17:13.097 } 00:17:13.097 Got JSON-RPC error response 00:17:13.097 response: 00:17:13.097 { 00:17:13.097 "code": -22, 00:17:13.097 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:17:13.097 } 00:17:13.097 15:23:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:17:13.097 15:23:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@655 -- # es=1 00:17:13.097 15:23:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:17:13.097 15:23:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:17:13.097 15:23:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:17:13.097 15:23:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@779 -- # sleep 1 00:17:14.038 15:23:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:14.038 15:23:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:14.038 15:23:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:14.038 15:23:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:14.038 15:23:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:14.038 15:23:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:14.038 15:23:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:14.038 15:23:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:14.038 15:23:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:14.038 15:23:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:14.038 15:23:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:14.038 15:23:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:14.038 15:23:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:14.038 15:23:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:14.038 15:23:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:14.038 15:23:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:14.038 "name": "raid_bdev1", 00:17:14.038 "uuid": "a49c1027-9619-4136-b2c0-3709f253da7b", 00:17:14.038 "strip_size_kb": 0, 00:17:14.038 "state": "online", 00:17:14.038 "raid_level": "raid1", 00:17:14.038 "superblock": true, 00:17:14.038 "num_base_bdevs": 2, 00:17:14.038 "num_base_bdevs_discovered": 1, 00:17:14.038 "num_base_bdevs_operational": 1, 00:17:14.038 "base_bdevs_list": [ 00:17:14.038 { 00:17:14.038 "name": null, 00:17:14.038 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:14.038 "is_configured": false, 00:17:14.038 "data_offset": 0, 00:17:14.038 "data_size": 7936 00:17:14.038 }, 00:17:14.038 { 00:17:14.038 "name": "BaseBdev2", 00:17:14.038 "uuid": "946223a0-8046-550a-94d0-a2cfe3831db5", 00:17:14.038 "is_configured": true, 00:17:14.038 "data_offset": 256, 00:17:14.038 "data_size": 7936 00:17:14.038 } 00:17:14.038 ] 00:17:14.038 }' 00:17:14.038 15:23:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:14.038 15:23:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:14.609 15:23:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:14.609 15:23:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:14.609 15:23:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:14.609 15:23:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:14.609 15:23:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:14.609 15:23:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:14.609 15:23:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:14.609 15:23:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:14.609 15:23:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:14.609 15:23:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:14.609 15:23:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:14.609 "name": "raid_bdev1", 00:17:14.609 "uuid": "a49c1027-9619-4136-b2c0-3709f253da7b", 00:17:14.609 "strip_size_kb": 0, 00:17:14.609 "state": "online", 00:17:14.609 "raid_level": "raid1", 00:17:14.609 "superblock": true, 00:17:14.609 "num_base_bdevs": 2, 00:17:14.609 "num_base_bdevs_discovered": 1, 00:17:14.609 "num_base_bdevs_operational": 1, 00:17:14.609 "base_bdevs_list": [ 00:17:14.609 { 00:17:14.609 "name": null, 00:17:14.609 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:14.609 "is_configured": false, 00:17:14.609 "data_offset": 0, 00:17:14.609 "data_size": 7936 00:17:14.609 }, 00:17:14.609 { 00:17:14.609 "name": "BaseBdev2", 00:17:14.609 "uuid": "946223a0-8046-550a-94d0-a2cfe3831db5", 00:17:14.609 "is_configured": true, 00:17:14.609 "data_offset": 256, 00:17:14.609 "data_size": 7936 00:17:14.609 } 00:17:14.609 ] 00:17:14.609 }' 00:17:14.609 15:23:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:14.609 15:23:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:14.609 15:23:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:14.609 15:23:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:14.609 15:23:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@784 -- # killprocess 96874 00:17:14.609 15:23:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@954 -- # '[' -z 96874 ']' 00:17:14.609 15:23:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@958 -- # kill -0 96874 00:17:14.609 15:23:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@959 -- # uname 00:17:14.609 15:23:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:17:14.609 15:23:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 96874 00:17:14.609 15:23:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:17:14.609 killing process with pid 96874 00:17:14.609 Received shutdown signal, test time was about 60.000000 seconds 00:17:14.609 00:17:14.609 Latency(us) 00:17:14.609 [2024-11-19T15:23:04.948Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:17:14.609 [2024-11-19T15:23:04.948Z] =================================================================================================================== 00:17:14.609 [2024-11-19T15:23:04.948Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:17:14.609 15:23:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:17:14.609 15:23:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@972 -- # echo 'killing process with pid 96874' 00:17:14.609 15:23:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@973 -- # kill 96874 00:17:14.609 [2024-11-19 15:23:04.919637] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:17:14.609 [2024-11-19 15:23:04.919739] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:14.609 [2024-11-19 15:23:04.919785] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:14.609 [2024-11-19 15:23:04.919794] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:17:14.609 15:23:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@978 -- # wait 96874 00:17:14.870 [2024-11-19 15:23:04.950975] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:17:14.870 15:23:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@786 -- # return 0 00:17:14.870 00:17:14.870 real 0m18.470s 00:17:14.870 user 0m24.538s 00:17:14.870 sys 0m2.710s 00:17:14.870 ************************************ 00:17:14.870 END TEST raid_rebuild_test_sb_4k 00:17:14.870 ************************************ 00:17:14.870 15:23:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@1130 -- # xtrace_disable 00:17:14.870 15:23:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:17:15.131 15:23:05 bdev_raid -- bdev/bdev_raid.sh@1003 -- # base_malloc_params='-m 32' 00:17:15.131 15:23:05 bdev_raid -- bdev/bdev_raid.sh@1004 -- # run_test raid_state_function_test_sb_md_separate raid_state_function_test raid1 2 true 00:17:15.131 15:23:05 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:17:15.131 15:23:05 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:17:15.131 15:23:05 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:17:15.131 ************************************ 00:17:15.131 START TEST raid_state_function_test_sb_md_separate 00:17:15.131 ************************************ 00:17:15.131 15:23:05 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 2 true 00:17:15.131 15:23:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:17:15.131 15:23:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:17:15.131 15:23:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:17:15.131 15:23:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:17:15.131 15:23:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:17:15.131 15:23:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:17:15.131 15:23:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:17:15.131 15:23:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:17:15.131 15:23:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:17:15.131 15:23:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:17:15.131 15:23:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:17:15.131 15:23:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:17:15.131 15:23:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:17:15.131 15:23:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:17:15.131 15:23:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:17:15.131 15:23:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@211 -- # local strip_size 00:17:15.131 15:23:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:17:15.131 15:23:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:17:15.131 15:23:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:17:15.131 15:23:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:17:15.131 15:23:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:17:15.131 15:23:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:17:15.131 15:23:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@229 -- # raid_pid=97551 00:17:15.131 15:23:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:17:15.131 15:23:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 97551' 00:17:15.131 Process raid pid: 97551 00:17:15.131 15:23:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@231 -- # waitforlisten 97551 00:17:15.131 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:17:15.131 15:23:05 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@835 -- # '[' -z 97551 ']' 00:17:15.131 15:23:05 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:17:15.131 15:23:05 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@840 -- # local max_retries=100 00:17:15.132 15:23:05 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:17:15.132 15:23:05 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@844 -- # xtrace_disable 00:17:15.132 15:23:05 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:15.132 [2024-11-19 15:23:05.325308] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:17:15.132 [2024-11-19 15:23:05.325525] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:17:15.392 [2024-11-19 15:23:05.483372] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:15.392 [2024-11-19 15:23:05.508826] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:15.392 [2024-11-19 15:23:05.551663] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:17:15.392 [2024-11-19 15:23:05.551783] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:17:15.962 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:17:15.962 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@868 -- # return 0 00:17:15.962 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:17:15.962 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:15.962 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:15.962 [2024-11-19 15:23:06.137477] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:17:15.962 [2024-11-19 15:23:06.137589] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:17:15.962 [2024-11-19 15:23:06.137632] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:17:15.962 [2024-11-19 15:23:06.137656] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:17:15.962 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:15.962 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:17:15.962 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:17:15.962 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:17:15.962 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:15.962 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:15.962 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:15.962 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:15.962 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:15.963 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:15.963 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:15.963 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:15.963 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:17:15.963 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:15.963 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:15.963 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:15.963 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:15.963 "name": "Existed_Raid", 00:17:15.963 "uuid": "28b5b503-2bb4-48de-aa2f-55a81f334cf8", 00:17:15.963 "strip_size_kb": 0, 00:17:15.963 "state": "configuring", 00:17:15.963 "raid_level": "raid1", 00:17:15.963 "superblock": true, 00:17:15.963 "num_base_bdevs": 2, 00:17:15.963 "num_base_bdevs_discovered": 0, 00:17:15.963 "num_base_bdevs_operational": 2, 00:17:15.963 "base_bdevs_list": [ 00:17:15.963 { 00:17:15.963 "name": "BaseBdev1", 00:17:15.963 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:15.963 "is_configured": false, 00:17:15.963 "data_offset": 0, 00:17:15.963 "data_size": 0 00:17:15.963 }, 00:17:15.963 { 00:17:15.963 "name": "BaseBdev2", 00:17:15.963 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:15.963 "is_configured": false, 00:17:15.963 "data_offset": 0, 00:17:15.963 "data_size": 0 00:17:15.963 } 00:17:15.963 ] 00:17:15.963 }' 00:17:15.963 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:15.963 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:16.533 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:17:16.533 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:16.533 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:16.533 [2024-11-19 15:23:06.624540] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:17:16.533 [2024-11-19 15:23:06.624616] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:17:16.533 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:16.533 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:17:16.533 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:16.533 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:16.533 [2024-11-19 15:23:06.636526] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:17:16.533 [2024-11-19 15:23:06.636613] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:17:16.533 [2024-11-19 15:23:06.636624] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:17:16.533 [2024-11-19 15:23:06.636644] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:17:16.533 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:16.533 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev1 00:17:16.533 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:16.533 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:16.533 [2024-11-19 15:23:06.657874] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:17:16.533 BaseBdev1 00:17:16.533 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:16.533 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:17:16.533 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:17:16.534 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:17:16.534 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@905 -- # local i 00:17:16.534 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:17:16.534 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:17:16.534 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:17:16.534 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:16.534 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:16.534 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:16.534 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:17:16.534 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:16.534 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:16.534 [ 00:17:16.534 { 00:17:16.534 "name": "BaseBdev1", 00:17:16.534 "aliases": [ 00:17:16.534 "7fbbc75b-40f5-42d9-abe6-18b1a470698e" 00:17:16.534 ], 00:17:16.534 "product_name": "Malloc disk", 00:17:16.534 "block_size": 4096, 00:17:16.534 "num_blocks": 8192, 00:17:16.534 "uuid": "7fbbc75b-40f5-42d9-abe6-18b1a470698e", 00:17:16.534 "md_size": 32, 00:17:16.534 "md_interleave": false, 00:17:16.534 "dif_type": 0, 00:17:16.534 "assigned_rate_limits": { 00:17:16.534 "rw_ios_per_sec": 0, 00:17:16.534 "rw_mbytes_per_sec": 0, 00:17:16.534 "r_mbytes_per_sec": 0, 00:17:16.534 "w_mbytes_per_sec": 0 00:17:16.534 }, 00:17:16.534 "claimed": true, 00:17:16.534 "claim_type": "exclusive_write", 00:17:16.534 "zoned": false, 00:17:16.534 "supported_io_types": { 00:17:16.534 "read": true, 00:17:16.534 "write": true, 00:17:16.534 "unmap": true, 00:17:16.534 "flush": true, 00:17:16.534 "reset": true, 00:17:16.534 "nvme_admin": false, 00:17:16.534 "nvme_io": false, 00:17:16.534 "nvme_io_md": false, 00:17:16.534 "write_zeroes": true, 00:17:16.534 "zcopy": true, 00:17:16.534 "get_zone_info": false, 00:17:16.534 "zone_management": false, 00:17:16.534 "zone_append": false, 00:17:16.534 "compare": false, 00:17:16.534 "compare_and_write": false, 00:17:16.534 "abort": true, 00:17:16.534 "seek_hole": false, 00:17:16.534 "seek_data": false, 00:17:16.534 "copy": true, 00:17:16.534 "nvme_iov_md": false 00:17:16.534 }, 00:17:16.534 "memory_domains": [ 00:17:16.534 { 00:17:16.534 "dma_device_id": "system", 00:17:16.534 "dma_device_type": 1 00:17:16.534 }, 00:17:16.534 { 00:17:16.534 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:16.534 "dma_device_type": 2 00:17:16.534 } 00:17:16.534 ], 00:17:16.534 "driver_specific": {} 00:17:16.534 } 00:17:16.534 ] 00:17:16.534 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:16.534 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@911 -- # return 0 00:17:16.534 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:17:16.534 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:17:16.534 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:17:16.534 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:16.534 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:16.534 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:16.534 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:16.534 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:16.534 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:16.534 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:16.534 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:17:16.534 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:16.534 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:16.534 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:16.534 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:16.534 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:16.534 "name": "Existed_Raid", 00:17:16.534 "uuid": "e2a181b3-fbb3-4ac1-aeb2-895f67818ac4", 00:17:16.534 "strip_size_kb": 0, 00:17:16.534 "state": "configuring", 00:17:16.534 "raid_level": "raid1", 00:17:16.534 "superblock": true, 00:17:16.534 "num_base_bdevs": 2, 00:17:16.534 "num_base_bdevs_discovered": 1, 00:17:16.534 "num_base_bdevs_operational": 2, 00:17:16.534 "base_bdevs_list": [ 00:17:16.534 { 00:17:16.534 "name": "BaseBdev1", 00:17:16.534 "uuid": "7fbbc75b-40f5-42d9-abe6-18b1a470698e", 00:17:16.534 "is_configured": true, 00:17:16.534 "data_offset": 256, 00:17:16.534 "data_size": 7936 00:17:16.534 }, 00:17:16.534 { 00:17:16.534 "name": "BaseBdev2", 00:17:16.534 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:16.534 "is_configured": false, 00:17:16.534 "data_offset": 0, 00:17:16.534 "data_size": 0 00:17:16.534 } 00:17:16.534 ] 00:17:16.534 }' 00:17:16.534 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:16.534 15:23:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:17.105 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:17:17.105 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:17.105 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:17.105 [2024-11-19 15:23:07.145081] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:17:17.105 [2024-11-19 15:23:07.145119] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:17:17.105 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:17.105 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:17:17.105 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:17.105 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:17.105 [2024-11-19 15:23:07.157109] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:17:17.105 [2024-11-19 15:23:07.158941] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:17:17.105 [2024-11-19 15:23:07.158991] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:17:17.105 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:17.105 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:17:17.105 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:17:17.105 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:17:17.105 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:17:17.105 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:17:17.105 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:17.105 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:17.105 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:17.105 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:17.105 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:17.105 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:17.105 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:17.105 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:17.105 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:17:17.105 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:17.105 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:17.106 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:17.106 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:17.106 "name": "Existed_Raid", 00:17:17.106 "uuid": "048c3757-5202-4b80-9816-dafffa66cda7", 00:17:17.106 "strip_size_kb": 0, 00:17:17.106 "state": "configuring", 00:17:17.106 "raid_level": "raid1", 00:17:17.106 "superblock": true, 00:17:17.106 "num_base_bdevs": 2, 00:17:17.106 "num_base_bdevs_discovered": 1, 00:17:17.106 "num_base_bdevs_operational": 2, 00:17:17.106 "base_bdevs_list": [ 00:17:17.106 { 00:17:17.106 "name": "BaseBdev1", 00:17:17.106 "uuid": "7fbbc75b-40f5-42d9-abe6-18b1a470698e", 00:17:17.106 "is_configured": true, 00:17:17.106 "data_offset": 256, 00:17:17.106 "data_size": 7936 00:17:17.106 }, 00:17:17.106 { 00:17:17.106 "name": "BaseBdev2", 00:17:17.106 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:17.106 "is_configured": false, 00:17:17.106 "data_offset": 0, 00:17:17.106 "data_size": 0 00:17:17.106 } 00:17:17.106 ] 00:17:17.106 }' 00:17:17.106 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:17.106 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:17.366 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev2 00:17:17.366 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:17.366 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:17.366 [2024-11-19 15:23:07.651934] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:17:17.366 [2024-11-19 15:23:07.652121] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:17:17.367 [2024-11-19 15:23:07.652135] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:17:17.367 [2024-11-19 15:23:07.652222] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:17:17.367 [2024-11-19 15:23:07.652337] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:17:17.367 [2024-11-19 15:23:07.652352] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:17:17.367 [2024-11-19 15:23:07.652429] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:17.367 BaseBdev2 00:17:17.367 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:17.367 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:17:17.367 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:17:17.367 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:17:17.367 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@905 -- # local i 00:17:17.367 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:17:17.367 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:17:17.367 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:17:17.367 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:17.367 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:17.367 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:17.367 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:17:17.367 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:17.367 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:17.367 [ 00:17:17.367 { 00:17:17.367 "name": "BaseBdev2", 00:17:17.367 "aliases": [ 00:17:17.367 "378d2d0a-2132-4362-9502-5420330d6cfb" 00:17:17.367 ], 00:17:17.367 "product_name": "Malloc disk", 00:17:17.367 "block_size": 4096, 00:17:17.367 "num_blocks": 8192, 00:17:17.367 "uuid": "378d2d0a-2132-4362-9502-5420330d6cfb", 00:17:17.367 "md_size": 32, 00:17:17.367 "md_interleave": false, 00:17:17.367 "dif_type": 0, 00:17:17.367 "assigned_rate_limits": { 00:17:17.367 "rw_ios_per_sec": 0, 00:17:17.367 "rw_mbytes_per_sec": 0, 00:17:17.367 "r_mbytes_per_sec": 0, 00:17:17.367 "w_mbytes_per_sec": 0 00:17:17.367 }, 00:17:17.367 "claimed": true, 00:17:17.367 "claim_type": "exclusive_write", 00:17:17.367 "zoned": false, 00:17:17.367 "supported_io_types": { 00:17:17.367 "read": true, 00:17:17.367 "write": true, 00:17:17.367 "unmap": true, 00:17:17.367 "flush": true, 00:17:17.367 "reset": true, 00:17:17.367 "nvme_admin": false, 00:17:17.367 "nvme_io": false, 00:17:17.367 "nvme_io_md": false, 00:17:17.367 "write_zeroes": true, 00:17:17.367 "zcopy": true, 00:17:17.367 "get_zone_info": false, 00:17:17.367 "zone_management": false, 00:17:17.367 "zone_append": false, 00:17:17.367 "compare": false, 00:17:17.367 "compare_and_write": false, 00:17:17.367 "abort": true, 00:17:17.367 "seek_hole": false, 00:17:17.367 "seek_data": false, 00:17:17.367 "copy": true, 00:17:17.367 "nvme_iov_md": false 00:17:17.367 }, 00:17:17.367 "memory_domains": [ 00:17:17.367 { 00:17:17.367 "dma_device_id": "system", 00:17:17.367 "dma_device_type": 1 00:17:17.367 }, 00:17:17.367 { 00:17:17.367 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:17.367 "dma_device_type": 2 00:17:17.367 } 00:17:17.367 ], 00:17:17.367 "driver_specific": {} 00:17:17.367 } 00:17:17.367 ] 00:17:17.367 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:17.367 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@911 -- # return 0 00:17:17.367 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:17:17.367 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:17:17.367 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:17:17.367 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:17:17.367 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:17.367 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:17.367 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:17.367 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:17.367 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:17.367 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:17.367 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:17.367 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:17.367 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:17.367 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:17.367 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:17:17.367 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:17.627 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:17.627 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:17.627 "name": "Existed_Raid", 00:17:17.627 "uuid": "048c3757-5202-4b80-9816-dafffa66cda7", 00:17:17.627 "strip_size_kb": 0, 00:17:17.627 "state": "online", 00:17:17.627 "raid_level": "raid1", 00:17:17.627 "superblock": true, 00:17:17.627 "num_base_bdevs": 2, 00:17:17.627 "num_base_bdevs_discovered": 2, 00:17:17.627 "num_base_bdevs_operational": 2, 00:17:17.627 "base_bdevs_list": [ 00:17:17.627 { 00:17:17.627 "name": "BaseBdev1", 00:17:17.627 "uuid": "7fbbc75b-40f5-42d9-abe6-18b1a470698e", 00:17:17.627 "is_configured": true, 00:17:17.627 "data_offset": 256, 00:17:17.627 "data_size": 7936 00:17:17.627 }, 00:17:17.627 { 00:17:17.627 "name": "BaseBdev2", 00:17:17.627 "uuid": "378d2d0a-2132-4362-9502-5420330d6cfb", 00:17:17.627 "is_configured": true, 00:17:17.627 "data_offset": 256, 00:17:17.627 "data_size": 7936 00:17:17.627 } 00:17:17.627 ] 00:17:17.627 }' 00:17:17.627 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:17.628 15:23:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:17.888 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:17:17.888 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:17:17.888 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:17:17.888 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:17:17.888 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@184 -- # local name 00:17:17.888 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:17:17.888 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:17:17.888 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:17:17.888 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:17.888 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:17.888 [2024-11-19 15:23:08.119636] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:17:17.888 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:17.888 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:17:17.888 "name": "Existed_Raid", 00:17:17.888 "aliases": [ 00:17:17.888 "048c3757-5202-4b80-9816-dafffa66cda7" 00:17:17.888 ], 00:17:17.888 "product_name": "Raid Volume", 00:17:17.888 "block_size": 4096, 00:17:17.888 "num_blocks": 7936, 00:17:17.888 "uuid": "048c3757-5202-4b80-9816-dafffa66cda7", 00:17:17.888 "md_size": 32, 00:17:17.888 "md_interleave": false, 00:17:17.888 "dif_type": 0, 00:17:17.888 "assigned_rate_limits": { 00:17:17.888 "rw_ios_per_sec": 0, 00:17:17.888 "rw_mbytes_per_sec": 0, 00:17:17.888 "r_mbytes_per_sec": 0, 00:17:17.888 "w_mbytes_per_sec": 0 00:17:17.888 }, 00:17:17.888 "claimed": false, 00:17:17.888 "zoned": false, 00:17:17.888 "supported_io_types": { 00:17:17.888 "read": true, 00:17:17.888 "write": true, 00:17:17.888 "unmap": false, 00:17:17.888 "flush": false, 00:17:17.888 "reset": true, 00:17:17.888 "nvme_admin": false, 00:17:17.888 "nvme_io": false, 00:17:17.888 "nvme_io_md": false, 00:17:17.888 "write_zeroes": true, 00:17:17.888 "zcopy": false, 00:17:17.888 "get_zone_info": false, 00:17:17.888 "zone_management": false, 00:17:17.889 "zone_append": false, 00:17:17.889 "compare": false, 00:17:17.889 "compare_and_write": false, 00:17:17.889 "abort": false, 00:17:17.889 "seek_hole": false, 00:17:17.889 "seek_data": false, 00:17:17.889 "copy": false, 00:17:17.889 "nvme_iov_md": false 00:17:17.889 }, 00:17:17.889 "memory_domains": [ 00:17:17.889 { 00:17:17.889 "dma_device_id": "system", 00:17:17.889 "dma_device_type": 1 00:17:17.889 }, 00:17:17.889 { 00:17:17.889 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:17.889 "dma_device_type": 2 00:17:17.889 }, 00:17:17.889 { 00:17:17.889 "dma_device_id": "system", 00:17:17.889 "dma_device_type": 1 00:17:17.889 }, 00:17:17.889 { 00:17:17.889 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:17.889 "dma_device_type": 2 00:17:17.889 } 00:17:17.889 ], 00:17:17.889 "driver_specific": { 00:17:17.889 "raid": { 00:17:17.889 "uuid": "048c3757-5202-4b80-9816-dafffa66cda7", 00:17:17.889 "strip_size_kb": 0, 00:17:17.889 "state": "online", 00:17:17.889 "raid_level": "raid1", 00:17:17.889 "superblock": true, 00:17:17.889 "num_base_bdevs": 2, 00:17:17.889 "num_base_bdevs_discovered": 2, 00:17:17.889 "num_base_bdevs_operational": 2, 00:17:17.889 "base_bdevs_list": [ 00:17:17.889 { 00:17:17.889 "name": "BaseBdev1", 00:17:17.889 "uuid": "7fbbc75b-40f5-42d9-abe6-18b1a470698e", 00:17:17.889 "is_configured": true, 00:17:17.889 "data_offset": 256, 00:17:17.889 "data_size": 7936 00:17:17.889 }, 00:17:17.889 { 00:17:17.889 "name": "BaseBdev2", 00:17:17.889 "uuid": "378d2d0a-2132-4362-9502-5420330d6cfb", 00:17:17.889 "is_configured": true, 00:17:17.889 "data_offset": 256, 00:17:17.889 "data_size": 7936 00:17:17.889 } 00:17:17.889 ] 00:17:17.889 } 00:17:17.889 } 00:17:17.889 }' 00:17:17.889 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:17:17.889 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:17:17.889 BaseBdev2' 00:17:17.889 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:17:18.149 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 32 false 0' 00:17:18.149 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:17:18.149 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:17:18.149 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:17:18.149 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:18.149 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:18.149 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:18.149 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:17:18.149 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:17:18.149 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:17:18.149 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:17:18.149 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:17:18.149 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:18.149 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:18.149 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:18.149 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:17:18.149 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:17:18.149 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:17:18.149 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:18.149 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:18.149 [2024-11-19 15:23:08.327110] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:17:18.149 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:18.149 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@260 -- # local expected_state 00:17:18.149 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:17:18.149 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@198 -- # case $1 in 00:17:18.149 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@199 -- # return 0 00:17:18.149 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:17:18.149 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:17:18.149 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:17:18.149 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:18.149 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:18.149 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:18.149 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:18.149 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:18.149 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:18.149 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:18.150 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:18.150 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:18.150 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:18.150 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:17:18.150 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:18.150 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:18.150 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:18.150 "name": "Existed_Raid", 00:17:18.150 "uuid": "048c3757-5202-4b80-9816-dafffa66cda7", 00:17:18.150 "strip_size_kb": 0, 00:17:18.150 "state": "online", 00:17:18.150 "raid_level": "raid1", 00:17:18.150 "superblock": true, 00:17:18.150 "num_base_bdevs": 2, 00:17:18.150 "num_base_bdevs_discovered": 1, 00:17:18.150 "num_base_bdevs_operational": 1, 00:17:18.150 "base_bdevs_list": [ 00:17:18.150 { 00:17:18.150 "name": null, 00:17:18.150 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:18.150 "is_configured": false, 00:17:18.150 "data_offset": 0, 00:17:18.150 "data_size": 7936 00:17:18.150 }, 00:17:18.150 { 00:17:18.150 "name": "BaseBdev2", 00:17:18.150 "uuid": "378d2d0a-2132-4362-9502-5420330d6cfb", 00:17:18.150 "is_configured": true, 00:17:18.150 "data_offset": 256, 00:17:18.150 "data_size": 7936 00:17:18.150 } 00:17:18.150 ] 00:17:18.150 }' 00:17:18.150 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:18.150 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:18.719 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:17:18.719 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:17:18.719 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:18.719 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:17:18.719 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:18.719 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:18.719 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:18.719 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:17:18.719 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:17:18.719 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:17:18.720 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:18.720 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:18.720 [2024-11-19 15:23:08.822423] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:17:18.720 [2024-11-19 15:23:08.822580] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:17:18.720 [2024-11-19 15:23:08.834865] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:18.720 [2024-11-19 15:23:08.834989] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:18.720 [2024-11-19 15:23:08.835033] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:17:18.720 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:18.720 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:17:18.720 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:17:18.720 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:17:18.720 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:18.720 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:18.720 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:18.720 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:18.720 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:17:18.720 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:17:18.720 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:17:18.720 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@326 -- # killprocess 97551 00:17:18.720 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@954 -- # '[' -z 97551 ']' 00:17:18.720 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@958 -- # kill -0 97551 00:17:18.720 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@959 -- # uname 00:17:18.720 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:17:18.720 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 97551 00:17:18.720 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:17:18.720 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:17:18.720 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@972 -- # echo 'killing process with pid 97551' 00:17:18.720 killing process with pid 97551 00:17:18.720 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@973 -- # kill 97551 00:17:18.720 [2024-11-19 15:23:08.930229] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:17:18.720 15:23:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@978 -- # wait 97551 00:17:18.720 [2024-11-19 15:23:08.931241] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:17:18.980 15:23:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@328 -- # return 0 00:17:18.980 00:17:18.980 real 0m3.919s 00:17:18.980 user 0m6.153s 00:17:18.980 sys 0m0.877s 00:17:18.980 15:23:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@1130 -- # xtrace_disable 00:17:18.980 15:23:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:18.980 ************************************ 00:17:18.980 END TEST raid_state_function_test_sb_md_separate 00:17:18.980 ************************************ 00:17:18.980 15:23:09 bdev_raid -- bdev/bdev_raid.sh@1005 -- # run_test raid_superblock_test_md_separate raid_superblock_test raid1 2 00:17:18.980 15:23:09 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:17:18.980 15:23:09 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:17:18.980 15:23:09 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:17:18.980 ************************************ 00:17:18.980 START TEST raid_superblock_test_md_separate 00:17:18.980 ************************************ 00:17:18.980 15:23:09 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@1129 -- # raid_superblock_test raid1 2 00:17:18.980 15:23:09 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:17:18.980 15:23:09 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:17:18.980 15:23:09 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:17:18.981 15:23:09 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:17:18.981 15:23:09 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:17:18.981 15:23:09 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:17:18.981 15:23:09 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:17:18.981 15:23:09 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:17:18.981 15:23:09 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:17:18.981 15:23:09 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@399 -- # local strip_size 00:17:18.981 15:23:09 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:17:18.981 15:23:09 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:17:18.981 15:23:09 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:17:18.981 15:23:09 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:17:18.981 15:23:09 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:17:18.981 15:23:09 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@412 -- # raid_pid=97792 00:17:18.981 15:23:09 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:17:18.981 15:23:09 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@413 -- # waitforlisten 97792 00:17:18.981 15:23:09 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@835 -- # '[' -z 97792 ']' 00:17:18.981 15:23:09 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:17:18.981 15:23:09 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@840 -- # local max_retries=100 00:17:18.981 15:23:09 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:17:18.981 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:17:18.981 15:23:09 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@844 -- # xtrace_disable 00:17:18.981 15:23:09 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:19.240 [2024-11-19 15:23:09.325362] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:17:19.241 [2024-11-19 15:23:09.326031] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid97792 ] 00:17:19.241 [2024-11-19 15:23:09.482005] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:19.241 [2024-11-19 15:23:09.506671] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:19.241 [2024-11-19 15:23:09.549921] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:17:19.241 [2024-11-19 15:23:09.550060] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:17:19.809 15:23:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:17:19.809 15:23:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@868 -- # return 0 00:17:19.809 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:17:19.809 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:17:19.809 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:17:19.809 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:17:19.809 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:17:19.809 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:17:19.809 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:17:19.809 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:17:19.809 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b malloc1 00:17:19.809 15:23:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:19.809 15:23:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:20.069 malloc1 00:17:20.069 15:23:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:20.069 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:17:20.069 15:23:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:20.069 15:23:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:20.069 [2024-11-19 15:23:10.161183] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:17:20.069 [2024-11-19 15:23:10.161325] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:20.069 [2024-11-19 15:23:10.161364] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:17:20.069 [2024-11-19 15:23:10.161397] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:20.069 [2024-11-19 15:23:10.163398] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:20.069 [2024-11-19 15:23:10.163470] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:17:20.069 pt1 00:17:20.069 15:23:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:20.069 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:17:20.069 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:17:20.069 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:17:20.069 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:17:20.069 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:17:20.069 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:17:20.069 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:17:20.069 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:17:20.069 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b malloc2 00:17:20.069 15:23:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:20.069 15:23:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:20.069 malloc2 00:17:20.069 15:23:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:20.069 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:17:20.069 15:23:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:20.069 15:23:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:20.069 [2024-11-19 15:23:10.190411] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:17:20.069 [2024-11-19 15:23:10.190461] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:20.069 [2024-11-19 15:23:10.190475] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:17:20.069 [2024-11-19 15:23:10.190485] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:20.069 [2024-11-19 15:23:10.192449] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:20.069 [2024-11-19 15:23:10.192539] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:17:20.069 pt2 00:17:20.069 15:23:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:20.069 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:17:20.069 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:17:20.069 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:17:20.069 15:23:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:20.069 15:23:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:20.069 [2024-11-19 15:23:10.202443] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:17:20.069 [2024-11-19 15:23:10.204408] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:17:20.069 [2024-11-19 15:23:10.204549] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:17:20.069 [2024-11-19 15:23:10.204565] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:17:20.069 [2024-11-19 15:23:10.204641] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:17:20.069 [2024-11-19 15:23:10.204735] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:17:20.069 [2024-11-19 15:23:10.204745] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:17:20.069 [2024-11-19 15:23:10.204842] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:20.069 15:23:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:20.069 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:17:20.069 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:20.069 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:20.069 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:20.069 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:20.069 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:20.069 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:20.069 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:20.069 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:20.069 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:20.069 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:20.070 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:20.070 15:23:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:20.070 15:23:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:20.070 15:23:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:20.070 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:20.070 "name": "raid_bdev1", 00:17:20.070 "uuid": "14a2e9f5-45ab-4590-b606-c9848942aba6", 00:17:20.070 "strip_size_kb": 0, 00:17:20.070 "state": "online", 00:17:20.070 "raid_level": "raid1", 00:17:20.070 "superblock": true, 00:17:20.070 "num_base_bdevs": 2, 00:17:20.070 "num_base_bdevs_discovered": 2, 00:17:20.070 "num_base_bdevs_operational": 2, 00:17:20.070 "base_bdevs_list": [ 00:17:20.070 { 00:17:20.070 "name": "pt1", 00:17:20.070 "uuid": "00000000-0000-0000-0000-000000000001", 00:17:20.070 "is_configured": true, 00:17:20.070 "data_offset": 256, 00:17:20.070 "data_size": 7936 00:17:20.070 }, 00:17:20.070 { 00:17:20.070 "name": "pt2", 00:17:20.070 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:20.070 "is_configured": true, 00:17:20.070 "data_offset": 256, 00:17:20.070 "data_size": 7936 00:17:20.070 } 00:17:20.070 ] 00:17:20.070 }' 00:17:20.070 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:20.070 15:23:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:20.330 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:17:20.330 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:17:20.330 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:17:20.330 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:17:20.330 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@184 -- # local name 00:17:20.330 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:17:20.330 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:17:20.330 15:23:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:20.330 15:23:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:20.330 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:17:20.330 [2024-11-19 15:23:10.657885] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:17:20.330 15:23:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:20.591 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:17:20.591 "name": "raid_bdev1", 00:17:20.591 "aliases": [ 00:17:20.591 "14a2e9f5-45ab-4590-b606-c9848942aba6" 00:17:20.591 ], 00:17:20.591 "product_name": "Raid Volume", 00:17:20.591 "block_size": 4096, 00:17:20.591 "num_blocks": 7936, 00:17:20.591 "uuid": "14a2e9f5-45ab-4590-b606-c9848942aba6", 00:17:20.591 "md_size": 32, 00:17:20.591 "md_interleave": false, 00:17:20.591 "dif_type": 0, 00:17:20.591 "assigned_rate_limits": { 00:17:20.591 "rw_ios_per_sec": 0, 00:17:20.591 "rw_mbytes_per_sec": 0, 00:17:20.591 "r_mbytes_per_sec": 0, 00:17:20.591 "w_mbytes_per_sec": 0 00:17:20.591 }, 00:17:20.591 "claimed": false, 00:17:20.591 "zoned": false, 00:17:20.591 "supported_io_types": { 00:17:20.591 "read": true, 00:17:20.591 "write": true, 00:17:20.591 "unmap": false, 00:17:20.591 "flush": false, 00:17:20.591 "reset": true, 00:17:20.591 "nvme_admin": false, 00:17:20.591 "nvme_io": false, 00:17:20.591 "nvme_io_md": false, 00:17:20.591 "write_zeroes": true, 00:17:20.591 "zcopy": false, 00:17:20.591 "get_zone_info": false, 00:17:20.591 "zone_management": false, 00:17:20.591 "zone_append": false, 00:17:20.591 "compare": false, 00:17:20.591 "compare_and_write": false, 00:17:20.591 "abort": false, 00:17:20.591 "seek_hole": false, 00:17:20.591 "seek_data": false, 00:17:20.591 "copy": false, 00:17:20.591 "nvme_iov_md": false 00:17:20.591 }, 00:17:20.591 "memory_domains": [ 00:17:20.591 { 00:17:20.591 "dma_device_id": "system", 00:17:20.591 "dma_device_type": 1 00:17:20.591 }, 00:17:20.591 { 00:17:20.591 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:20.591 "dma_device_type": 2 00:17:20.591 }, 00:17:20.591 { 00:17:20.591 "dma_device_id": "system", 00:17:20.591 "dma_device_type": 1 00:17:20.591 }, 00:17:20.591 { 00:17:20.591 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:20.591 "dma_device_type": 2 00:17:20.591 } 00:17:20.591 ], 00:17:20.591 "driver_specific": { 00:17:20.591 "raid": { 00:17:20.591 "uuid": "14a2e9f5-45ab-4590-b606-c9848942aba6", 00:17:20.591 "strip_size_kb": 0, 00:17:20.591 "state": "online", 00:17:20.591 "raid_level": "raid1", 00:17:20.591 "superblock": true, 00:17:20.591 "num_base_bdevs": 2, 00:17:20.591 "num_base_bdevs_discovered": 2, 00:17:20.591 "num_base_bdevs_operational": 2, 00:17:20.591 "base_bdevs_list": [ 00:17:20.591 { 00:17:20.591 "name": "pt1", 00:17:20.591 "uuid": "00000000-0000-0000-0000-000000000001", 00:17:20.591 "is_configured": true, 00:17:20.591 "data_offset": 256, 00:17:20.591 "data_size": 7936 00:17:20.591 }, 00:17:20.591 { 00:17:20.591 "name": "pt2", 00:17:20.591 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:20.591 "is_configured": true, 00:17:20.591 "data_offset": 256, 00:17:20.591 "data_size": 7936 00:17:20.591 } 00:17:20.591 ] 00:17:20.591 } 00:17:20.591 } 00:17:20.591 }' 00:17:20.591 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:17:20.591 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:17:20.591 pt2' 00:17:20.591 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:17:20.591 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 32 false 0' 00:17:20.591 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:17:20.591 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:17:20.591 15:23:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:20.591 15:23:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:20.591 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:17:20.591 15:23:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:20.591 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:17:20.591 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:17:20.591 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:17:20.591 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:17:20.591 15:23:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:20.591 15:23:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:20.591 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:17:20.591 15:23:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:20.591 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:17:20.591 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:17:20.591 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:17:20.591 15:23:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:20.591 15:23:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:20.591 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:17:20.591 [2024-11-19 15:23:10.909356] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:17:20.852 15:23:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:20.852 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=14a2e9f5-45ab-4590-b606-c9848942aba6 00:17:20.852 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@436 -- # '[' -z 14a2e9f5-45ab-4590-b606-c9848942aba6 ']' 00:17:20.852 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:17:20.852 15:23:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:20.852 15:23:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:20.852 [2024-11-19 15:23:10.957083] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:17:20.852 [2024-11-19 15:23:10.957160] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:17:20.852 [2024-11-19 15:23:10.957262] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:20.852 [2024-11-19 15:23:10.957335] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:20.852 [2024-11-19 15:23:10.957384] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:17:20.852 15:23:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:20.852 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:20.852 15:23:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:20.852 15:23:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:20.852 15:23:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:17:20.852 15:23:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:20.852 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:17:20.852 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:17:20.852 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@652 -- # local es=0 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:20.853 [2024-11-19 15:23:11.092852] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:17:20.853 [2024-11-19 15:23:11.094709] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:17:20.853 [2024-11-19 15:23:11.094820] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:17:20.853 [2024-11-19 15:23:11.094897] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:17:20.853 [2024-11-19 15:23:11.094962] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:17:20.853 [2024-11-19 15:23:11.094995] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:17:20.853 request: 00:17:20.853 { 00:17:20.853 "name": "raid_bdev1", 00:17:20.853 "raid_level": "raid1", 00:17:20.853 "base_bdevs": [ 00:17:20.853 "malloc1", 00:17:20.853 "malloc2" 00:17:20.853 ], 00:17:20.853 "superblock": false, 00:17:20.853 "method": "bdev_raid_create", 00:17:20.853 "req_id": 1 00:17:20.853 } 00:17:20.853 Got JSON-RPC error response 00:17:20.853 response: 00:17:20.853 { 00:17:20.853 "code": -17, 00:17:20.853 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:17:20.853 } 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@655 -- # es=1 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:20.853 [2024-11-19 15:23:11.156704] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:17:20.853 [2024-11-19 15:23:11.156815] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:20.853 [2024-11-19 15:23:11.156850] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:17:20.853 [2024-11-19 15:23:11.156876] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:20.853 [2024-11-19 15:23:11.158773] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:20.853 [2024-11-19 15:23:11.158854] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:17:20.853 [2024-11-19 15:23:11.158912] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:17:20.853 [2024-11-19 15:23:11.158969] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:17:20.853 pt1 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:20.853 15:23:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:21.114 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:21.114 "name": "raid_bdev1", 00:17:21.114 "uuid": "14a2e9f5-45ab-4590-b606-c9848942aba6", 00:17:21.114 "strip_size_kb": 0, 00:17:21.114 "state": "configuring", 00:17:21.114 "raid_level": "raid1", 00:17:21.114 "superblock": true, 00:17:21.114 "num_base_bdevs": 2, 00:17:21.114 "num_base_bdevs_discovered": 1, 00:17:21.114 "num_base_bdevs_operational": 2, 00:17:21.114 "base_bdevs_list": [ 00:17:21.114 { 00:17:21.114 "name": "pt1", 00:17:21.114 "uuid": "00000000-0000-0000-0000-000000000001", 00:17:21.114 "is_configured": true, 00:17:21.114 "data_offset": 256, 00:17:21.114 "data_size": 7936 00:17:21.114 }, 00:17:21.114 { 00:17:21.114 "name": null, 00:17:21.114 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:21.114 "is_configured": false, 00:17:21.114 "data_offset": 256, 00:17:21.114 "data_size": 7936 00:17:21.114 } 00:17:21.114 ] 00:17:21.114 }' 00:17:21.114 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:21.114 15:23:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:21.374 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:17:21.374 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:17:21.374 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:17:21.374 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:17:21.374 15:23:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:21.374 15:23:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:21.374 [2024-11-19 15:23:11.611951] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:17:21.374 [2024-11-19 15:23:11.612079] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:21.374 [2024-11-19 15:23:11.612101] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:17:21.374 [2024-11-19 15:23:11.612109] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:21.374 [2024-11-19 15:23:11.612282] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:21.374 [2024-11-19 15:23:11.612296] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:17:21.374 [2024-11-19 15:23:11.612335] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:17:21.374 [2024-11-19 15:23:11.612353] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:17:21.374 [2024-11-19 15:23:11.612436] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:17:21.374 [2024-11-19 15:23:11.612445] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:17:21.374 [2024-11-19 15:23:11.612517] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:17:21.374 [2024-11-19 15:23:11.612596] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:17:21.374 [2024-11-19 15:23:11.612609] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:17:21.374 [2024-11-19 15:23:11.612667] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:21.374 pt2 00:17:21.374 15:23:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:21.374 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:17:21.374 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:17:21.374 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:17:21.374 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:21.374 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:21.374 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:21.374 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:21.374 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:21.374 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:21.374 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:21.374 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:21.374 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:21.374 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:21.374 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:21.374 15:23:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:21.374 15:23:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:21.375 15:23:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:21.375 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:21.375 "name": "raid_bdev1", 00:17:21.375 "uuid": "14a2e9f5-45ab-4590-b606-c9848942aba6", 00:17:21.375 "strip_size_kb": 0, 00:17:21.375 "state": "online", 00:17:21.375 "raid_level": "raid1", 00:17:21.375 "superblock": true, 00:17:21.375 "num_base_bdevs": 2, 00:17:21.375 "num_base_bdevs_discovered": 2, 00:17:21.375 "num_base_bdevs_operational": 2, 00:17:21.375 "base_bdevs_list": [ 00:17:21.375 { 00:17:21.375 "name": "pt1", 00:17:21.375 "uuid": "00000000-0000-0000-0000-000000000001", 00:17:21.375 "is_configured": true, 00:17:21.375 "data_offset": 256, 00:17:21.375 "data_size": 7936 00:17:21.375 }, 00:17:21.375 { 00:17:21.375 "name": "pt2", 00:17:21.375 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:21.375 "is_configured": true, 00:17:21.375 "data_offset": 256, 00:17:21.375 "data_size": 7936 00:17:21.375 } 00:17:21.375 ] 00:17:21.375 }' 00:17:21.375 15:23:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:21.375 15:23:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:21.946 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:17:21.946 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:17:21.946 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:17:21.946 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:17:21.946 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@184 -- # local name 00:17:21.946 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:17:21.946 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:17:21.946 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:17:21.946 15:23:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:21.946 15:23:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:21.946 [2024-11-19 15:23:12.071394] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:17:21.946 15:23:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:21.946 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:17:21.946 "name": "raid_bdev1", 00:17:21.946 "aliases": [ 00:17:21.946 "14a2e9f5-45ab-4590-b606-c9848942aba6" 00:17:21.946 ], 00:17:21.946 "product_name": "Raid Volume", 00:17:21.946 "block_size": 4096, 00:17:21.946 "num_blocks": 7936, 00:17:21.946 "uuid": "14a2e9f5-45ab-4590-b606-c9848942aba6", 00:17:21.946 "md_size": 32, 00:17:21.946 "md_interleave": false, 00:17:21.946 "dif_type": 0, 00:17:21.946 "assigned_rate_limits": { 00:17:21.946 "rw_ios_per_sec": 0, 00:17:21.946 "rw_mbytes_per_sec": 0, 00:17:21.946 "r_mbytes_per_sec": 0, 00:17:21.946 "w_mbytes_per_sec": 0 00:17:21.946 }, 00:17:21.946 "claimed": false, 00:17:21.946 "zoned": false, 00:17:21.946 "supported_io_types": { 00:17:21.946 "read": true, 00:17:21.946 "write": true, 00:17:21.946 "unmap": false, 00:17:21.946 "flush": false, 00:17:21.946 "reset": true, 00:17:21.946 "nvme_admin": false, 00:17:21.946 "nvme_io": false, 00:17:21.946 "nvme_io_md": false, 00:17:21.946 "write_zeroes": true, 00:17:21.946 "zcopy": false, 00:17:21.946 "get_zone_info": false, 00:17:21.946 "zone_management": false, 00:17:21.946 "zone_append": false, 00:17:21.946 "compare": false, 00:17:21.946 "compare_and_write": false, 00:17:21.946 "abort": false, 00:17:21.946 "seek_hole": false, 00:17:21.946 "seek_data": false, 00:17:21.946 "copy": false, 00:17:21.946 "nvme_iov_md": false 00:17:21.946 }, 00:17:21.946 "memory_domains": [ 00:17:21.946 { 00:17:21.946 "dma_device_id": "system", 00:17:21.946 "dma_device_type": 1 00:17:21.946 }, 00:17:21.946 { 00:17:21.946 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:21.946 "dma_device_type": 2 00:17:21.946 }, 00:17:21.946 { 00:17:21.946 "dma_device_id": "system", 00:17:21.946 "dma_device_type": 1 00:17:21.946 }, 00:17:21.946 { 00:17:21.946 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:21.946 "dma_device_type": 2 00:17:21.946 } 00:17:21.946 ], 00:17:21.946 "driver_specific": { 00:17:21.946 "raid": { 00:17:21.946 "uuid": "14a2e9f5-45ab-4590-b606-c9848942aba6", 00:17:21.946 "strip_size_kb": 0, 00:17:21.946 "state": "online", 00:17:21.946 "raid_level": "raid1", 00:17:21.946 "superblock": true, 00:17:21.946 "num_base_bdevs": 2, 00:17:21.946 "num_base_bdevs_discovered": 2, 00:17:21.946 "num_base_bdevs_operational": 2, 00:17:21.946 "base_bdevs_list": [ 00:17:21.946 { 00:17:21.946 "name": "pt1", 00:17:21.946 "uuid": "00000000-0000-0000-0000-000000000001", 00:17:21.946 "is_configured": true, 00:17:21.946 "data_offset": 256, 00:17:21.946 "data_size": 7936 00:17:21.946 }, 00:17:21.946 { 00:17:21.946 "name": "pt2", 00:17:21.946 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:21.946 "is_configured": true, 00:17:21.946 "data_offset": 256, 00:17:21.946 "data_size": 7936 00:17:21.946 } 00:17:21.946 ] 00:17:21.946 } 00:17:21.946 } 00:17:21.946 }' 00:17:21.946 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:17:21.946 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:17:21.946 pt2' 00:17:21.946 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:17:21.946 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 32 false 0' 00:17:21.946 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:17:21.946 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:17:21.946 15:23:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:21.946 15:23:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:21.946 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:17:21.946 15:23:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:21.946 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:17:21.946 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:17:21.946 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:17:21.946 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:17:21.946 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:17:21.946 15:23:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:21.946 15:23:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:22.207 15:23:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:22.207 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:17:22.207 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:17:22.207 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:17:22.207 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:17:22.207 15:23:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:22.207 15:23:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:22.207 [2024-11-19 15:23:12.326996] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:17:22.207 15:23:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:22.207 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@487 -- # '[' 14a2e9f5-45ab-4590-b606-c9848942aba6 '!=' 14a2e9f5-45ab-4590-b606-c9848942aba6 ']' 00:17:22.207 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:17:22.207 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@198 -- # case $1 in 00:17:22.207 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@199 -- # return 0 00:17:22.207 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:17:22.207 15:23:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:22.207 15:23:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:22.207 [2024-11-19 15:23:12.370722] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:17:22.207 15:23:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:22.207 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:22.207 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:22.207 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:22.207 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:22.207 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:22.207 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:22.207 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:22.207 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:22.207 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:22.207 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:22.207 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:22.207 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:22.207 15:23:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:22.207 15:23:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:22.207 15:23:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:22.207 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:22.207 "name": "raid_bdev1", 00:17:22.207 "uuid": "14a2e9f5-45ab-4590-b606-c9848942aba6", 00:17:22.207 "strip_size_kb": 0, 00:17:22.207 "state": "online", 00:17:22.207 "raid_level": "raid1", 00:17:22.207 "superblock": true, 00:17:22.207 "num_base_bdevs": 2, 00:17:22.207 "num_base_bdevs_discovered": 1, 00:17:22.207 "num_base_bdevs_operational": 1, 00:17:22.207 "base_bdevs_list": [ 00:17:22.207 { 00:17:22.207 "name": null, 00:17:22.207 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:22.207 "is_configured": false, 00:17:22.207 "data_offset": 0, 00:17:22.207 "data_size": 7936 00:17:22.207 }, 00:17:22.207 { 00:17:22.207 "name": "pt2", 00:17:22.207 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:22.207 "is_configured": true, 00:17:22.207 "data_offset": 256, 00:17:22.207 "data_size": 7936 00:17:22.207 } 00:17:22.207 ] 00:17:22.207 }' 00:17:22.207 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:22.207 15:23:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:22.467 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:17:22.467 15:23:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:22.467 15:23:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:22.467 [2024-11-19 15:23:12.778032] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:17:22.467 [2024-11-19 15:23:12.778056] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:17:22.467 [2024-11-19 15:23:12.778105] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:22.467 [2024-11-19 15:23:12.778143] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:22.467 [2024-11-19 15:23:12.778151] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:17:22.467 15:23:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:22.467 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:17:22.467 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:22.467 15:23:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:22.467 15:23:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:22.467 15:23:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:22.738 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:17:22.738 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:17:22.738 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:17:22.738 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:17:22.738 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:17:22.738 15:23:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:22.738 15:23:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:22.738 15:23:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:22.738 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:17:22.738 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:17:22.738 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:17:22.738 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:17:22.738 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@519 -- # i=1 00:17:22.738 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:17:22.738 15:23:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:22.738 15:23:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:22.738 [2024-11-19 15:23:12.849893] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:17:22.738 [2024-11-19 15:23:12.850018] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:22.738 [2024-11-19 15:23:12.850040] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:17:22.738 [2024-11-19 15:23:12.850048] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:22.738 [2024-11-19 15:23:12.851907] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:22.738 [2024-11-19 15:23:12.851944] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:17:22.738 [2024-11-19 15:23:12.851996] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:17:22.738 [2024-11-19 15:23:12.852031] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:17:22.738 [2024-11-19 15:23:12.852098] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:17:22.738 [2024-11-19 15:23:12.852106] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:17:22.738 [2024-11-19 15:23:12.852163] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:17:22.738 [2024-11-19 15:23:12.852238] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:17:22.739 [2024-11-19 15:23:12.852249] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:17:22.739 [2024-11-19 15:23:12.852306] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:22.739 pt2 00:17:22.739 15:23:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:22.739 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:22.739 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:22.739 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:22.739 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:22.739 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:22.739 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:22.739 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:22.739 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:22.739 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:22.739 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:22.739 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:22.739 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:22.739 15:23:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:22.739 15:23:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:22.739 15:23:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:22.739 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:22.739 "name": "raid_bdev1", 00:17:22.739 "uuid": "14a2e9f5-45ab-4590-b606-c9848942aba6", 00:17:22.739 "strip_size_kb": 0, 00:17:22.739 "state": "online", 00:17:22.739 "raid_level": "raid1", 00:17:22.739 "superblock": true, 00:17:22.739 "num_base_bdevs": 2, 00:17:22.739 "num_base_bdevs_discovered": 1, 00:17:22.739 "num_base_bdevs_operational": 1, 00:17:22.739 "base_bdevs_list": [ 00:17:22.739 { 00:17:22.739 "name": null, 00:17:22.739 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:22.739 "is_configured": false, 00:17:22.739 "data_offset": 256, 00:17:22.739 "data_size": 7936 00:17:22.739 }, 00:17:22.739 { 00:17:22.739 "name": "pt2", 00:17:22.739 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:22.739 "is_configured": true, 00:17:22.739 "data_offset": 256, 00:17:22.739 "data_size": 7936 00:17:22.739 } 00:17:22.739 ] 00:17:22.739 }' 00:17:22.739 15:23:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:22.739 15:23:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:23.021 15:23:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:17:23.021 15:23:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:23.021 15:23:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:23.021 [2024-11-19 15:23:13.329052] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:17:23.021 [2024-11-19 15:23:13.329125] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:17:23.021 [2024-11-19 15:23:13.329200] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:23.021 [2024-11-19 15:23:13.329252] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:23.021 [2024-11-19 15:23:13.329287] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:17:23.021 15:23:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:23.021 15:23:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:23.021 15:23:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:17:23.021 15:23:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:23.021 15:23:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:23.021 15:23:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:23.295 15:23:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:17:23.295 15:23:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:17:23.295 15:23:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:17:23.295 15:23:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:17:23.295 15:23:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:23.295 15:23:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:23.295 [2024-11-19 15:23:13.373033] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:17:23.295 [2024-11-19 15:23:13.373080] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:23.295 [2024-11-19 15:23:13.373097] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008d80 00:17:23.295 [2024-11-19 15:23:13.373109] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:23.295 [2024-11-19 15:23:13.375014] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:23.295 [2024-11-19 15:23:13.375086] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:17:23.295 [2024-11-19 15:23:13.375130] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:17:23.295 [2024-11-19 15:23:13.375159] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:17:23.295 [2024-11-19 15:23:13.375253] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:17:23.295 [2024-11-19 15:23:13.375266] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:17:23.295 [2024-11-19 15:23:13.375290] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:17:23.295 [2024-11-19 15:23:13.375320] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:17:23.295 [2024-11-19 15:23:13.375373] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:17:23.295 [2024-11-19 15:23:13.375383] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:17:23.295 [2024-11-19 15:23:13.375434] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:17:23.295 [2024-11-19 15:23:13.375499] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:17:23.295 [2024-11-19 15:23:13.375506] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:17:23.295 [2024-11-19 15:23:13.375580] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:23.295 pt1 00:17:23.295 15:23:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:23.295 15:23:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:17:23.295 15:23:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:23.295 15:23:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:23.295 15:23:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:23.295 15:23:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:23.295 15:23:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:23.295 15:23:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:23.295 15:23:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:23.295 15:23:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:23.295 15:23:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:23.295 15:23:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:23.295 15:23:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:23.295 15:23:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:23.295 15:23:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:23.295 15:23:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:23.295 15:23:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:23.295 15:23:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:23.295 "name": "raid_bdev1", 00:17:23.295 "uuid": "14a2e9f5-45ab-4590-b606-c9848942aba6", 00:17:23.295 "strip_size_kb": 0, 00:17:23.295 "state": "online", 00:17:23.295 "raid_level": "raid1", 00:17:23.295 "superblock": true, 00:17:23.295 "num_base_bdevs": 2, 00:17:23.295 "num_base_bdevs_discovered": 1, 00:17:23.295 "num_base_bdevs_operational": 1, 00:17:23.295 "base_bdevs_list": [ 00:17:23.295 { 00:17:23.295 "name": null, 00:17:23.295 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:23.295 "is_configured": false, 00:17:23.295 "data_offset": 256, 00:17:23.295 "data_size": 7936 00:17:23.295 }, 00:17:23.295 { 00:17:23.295 "name": "pt2", 00:17:23.295 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:23.295 "is_configured": true, 00:17:23.295 "data_offset": 256, 00:17:23.295 "data_size": 7936 00:17:23.295 } 00:17:23.295 ] 00:17:23.295 }' 00:17:23.295 15:23:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:23.295 15:23:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:23.556 15:23:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:17:23.556 15:23:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:17:23.556 15:23:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:23.556 15:23:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:23.556 15:23:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:23.556 15:23:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:17:23.556 15:23:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:17:23.556 15:23:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:23.556 15:23:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:23.556 15:23:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:17:23.556 [2024-11-19 15:23:13.832456] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:17:23.556 15:23:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:23.556 15:23:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@558 -- # '[' 14a2e9f5-45ab-4590-b606-c9848942aba6 '!=' 14a2e9f5-45ab-4590-b606-c9848942aba6 ']' 00:17:23.556 15:23:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@563 -- # killprocess 97792 00:17:23.556 15:23:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@954 -- # '[' -z 97792 ']' 00:17:23.556 15:23:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@958 -- # kill -0 97792 00:17:23.556 15:23:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@959 -- # uname 00:17:23.556 15:23:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:17:23.556 15:23:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 97792 00:17:23.816 15:23:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:17:23.816 15:23:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:17:23.816 15:23:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@972 -- # echo 'killing process with pid 97792' 00:17:23.816 killing process with pid 97792 00:17:23.816 15:23:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@973 -- # kill 97792 00:17:23.816 [2024-11-19 15:23:13.918549] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:17:23.816 [2024-11-19 15:23:13.918664] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:23.816 [2024-11-19 15:23:13.918732] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:23.816 [2024-11-19 15:23:13.918777] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:17:23.816 15:23:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@978 -- # wait 97792 00:17:23.816 [2024-11-19 15:23:13.942686] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:17:24.076 15:23:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@565 -- # return 0 00:17:24.076 00:17:24.076 real 0m4.926s 00:17:24.076 user 0m8.010s 00:17:24.076 sys 0m1.099s 00:17:24.076 15:23:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@1130 -- # xtrace_disable 00:17:24.076 15:23:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:24.076 ************************************ 00:17:24.076 END TEST raid_superblock_test_md_separate 00:17:24.076 ************************************ 00:17:24.076 15:23:14 bdev_raid -- bdev/bdev_raid.sh@1006 -- # '[' true = true ']' 00:17:24.076 15:23:14 bdev_raid -- bdev/bdev_raid.sh@1007 -- # run_test raid_rebuild_test_sb_md_separate raid_rebuild_test raid1 2 true false true 00:17:24.076 15:23:14 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:17:24.076 15:23:14 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:17:24.076 15:23:14 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:17:24.076 ************************************ 00:17:24.076 START TEST raid_rebuild_test_sb_md_separate 00:17:24.076 ************************************ 00:17:24.076 15:23:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 true false true 00:17:24.076 15:23:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:17:24.076 15:23:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:17:24.076 15:23:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:17:24.076 15:23:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:17:24.076 15:23:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@573 -- # local verify=true 00:17:24.076 15:23:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:17:24.076 15:23:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:17:24.076 15:23:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:17:24.076 15:23:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:17:24.076 15:23:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:17:24.076 15:23:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:17:24.076 15:23:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:17:24.076 15:23:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:17:24.076 15:23:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:17:24.076 15:23:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:17:24.076 15:23:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:17:24.076 15:23:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@576 -- # local strip_size 00:17:24.076 15:23:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@577 -- # local create_arg 00:17:24.076 15:23:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:17:24.076 15:23:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@579 -- # local data_offset 00:17:24.076 15:23:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:17:24.076 15:23:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:17:24.076 15:23:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:17:24.076 15:23:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:17:24.076 15:23:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@597 -- # raid_pid=98104 00:17:24.076 15:23:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:17:24.076 15:23:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@598 -- # waitforlisten 98104 00:17:24.076 15:23:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@835 -- # '[' -z 98104 ']' 00:17:24.076 15:23:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:17:24.076 15:23:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@840 -- # local max_retries=100 00:17:24.076 15:23:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:17:24.076 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:17:24.076 15:23:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@844 -- # xtrace_disable 00:17:24.076 15:23:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:24.076 [2024-11-19 15:23:14.344930] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:17:24.077 [2024-11-19 15:23:14.345635] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid98104 ] 00:17:24.077 I/O size of 3145728 is greater than zero copy threshold (65536). 00:17:24.077 Zero copy mechanism will not be used. 00:17:24.336 [2024-11-19 15:23:14.501502] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:24.337 [2024-11-19 15:23:14.525699] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:24.337 [2024-11-19 15:23:14.568714] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:17:24.337 [2024-11-19 15:23:14.568823] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:17:24.907 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:17:24.907 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@868 -- # return 0 00:17:24.907 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:17:24.907 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev1_malloc 00:17:24.907 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:24.907 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:24.907 BaseBdev1_malloc 00:17:24.907 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:24.907 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:17:24.907 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:24.907 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:24.907 [2024-11-19 15:23:15.183959] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:17:24.907 [2024-11-19 15:23:15.184147] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:24.907 [2024-11-19 15:23:15.184189] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:17:24.907 [2024-11-19 15:23:15.184233] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:24.907 [2024-11-19 15:23:15.186160] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:24.907 [2024-11-19 15:23:15.186242] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:17:24.907 BaseBdev1 00:17:24.907 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:24.907 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:17:24.907 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev2_malloc 00:17:24.907 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:24.907 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:24.907 BaseBdev2_malloc 00:17:24.907 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:24.907 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:17:24.907 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:24.907 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:24.907 [2024-11-19 15:23:15.213134] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:17:24.907 [2024-11-19 15:23:15.213244] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:24.907 [2024-11-19 15:23:15.213281] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:17:24.907 [2024-11-19 15:23:15.213307] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:24.907 [2024-11-19 15:23:15.215174] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:24.907 [2024-11-19 15:23:15.215254] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:17:24.907 BaseBdev2 00:17:24.907 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:24.907 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b spare_malloc 00:17:24.907 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:24.907 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:24.907 spare_malloc 00:17:25.168 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:25.168 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:17:25.168 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:25.168 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:25.168 spare_delay 00:17:25.168 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:25.168 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:17:25.168 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:25.168 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:25.168 [2024-11-19 15:23:15.263533] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:17:25.168 [2024-11-19 15:23:15.263583] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:25.168 [2024-11-19 15:23:15.263602] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:17:25.168 [2024-11-19 15:23:15.263610] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:25.168 [2024-11-19 15:23:15.265450] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:25.168 [2024-11-19 15:23:15.265485] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:17:25.168 spare 00:17:25.168 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:25.168 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:17:25.168 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:25.168 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:25.168 [2024-11-19 15:23:15.275553] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:17:25.168 [2024-11-19 15:23:15.277412] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:17:25.168 [2024-11-19 15:23:15.277556] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:17:25.169 [2024-11-19 15:23:15.277568] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:17:25.169 [2024-11-19 15:23:15.277664] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:17:25.169 [2024-11-19 15:23:15.277771] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:17:25.169 [2024-11-19 15:23:15.277782] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:17:25.169 [2024-11-19 15:23:15.277862] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:25.169 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:25.169 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:17:25.169 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:25.169 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:25.169 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:25.169 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:25.169 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:25.169 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:25.169 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:25.169 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:25.169 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:25.169 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:25.169 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:25.169 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:25.169 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:25.169 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:25.169 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:25.169 "name": "raid_bdev1", 00:17:25.169 "uuid": "24259e17-22de-45a2-b4ae-a059493f30d4", 00:17:25.169 "strip_size_kb": 0, 00:17:25.169 "state": "online", 00:17:25.169 "raid_level": "raid1", 00:17:25.169 "superblock": true, 00:17:25.169 "num_base_bdevs": 2, 00:17:25.169 "num_base_bdevs_discovered": 2, 00:17:25.169 "num_base_bdevs_operational": 2, 00:17:25.169 "base_bdevs_list": [ 00:17:25.169 { 00:17:25.169 "name": "BaseBdev1", 00:17:25.169 "uuid": "5f7c6115-3b26-5336-8d8a-2d1765850c0f", 00:17:25.169 "is_configured": true, 00:17:25.169 "data_offset": 256, 00:17:25.169 "data_size": 7936 00:17:25.169 }, 00:17:25.169 { 00:17:25.169 "name": "BaseBdev2", 00:17:25.169 "uuid": "38aec866-01f4-5eb5-9ba1-80c65063c764", 00:17:25.169 "is_configured": true, 00:17:25.169 "data_offset": 256, 00:17:25.169 "data_size": 7936 00:17:25.169 } 00:17:25.169 ] 00:17:25.169 }' 00:17:25.169 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:25.169 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:25.429 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:17:25.429 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:17:25.429 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:25.429 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:25.429 [2024-11-19 15:23:15.758976] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:17:25.690 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:25.690 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=7936 00:17:25.690 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:25.690 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:25.690 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:25.690 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:17:25.690 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:25.690 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@619 -- # data_offset=256 00:17:25.690 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:17:25.690 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:17:25.690 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:17:25.690 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:17:25.690 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:17:25.690 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:17:25.690 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # local bdev_list 00:17:25.690 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:17:25.690 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # local nbd_list 00:17:25.690 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@12 -- # local i 00:17:25.690 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:17:25.690 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:17:25.690 15:23:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:17:25.690 [2024-11-19 15:23:16.022291] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:17:25.950 /dev/nbd0 00:17:25.950 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:17:25.950 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:17:25.950 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:17:25.950 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@873 -- # local i 00:17:25.950 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:17:25.950 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:17:25.950 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:17:25.950 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@877 -- # break 00:17:25.950 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:17:25.950 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:17:25.950 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:17:25.950 1+0 records in 00:17:25.950 1+0 records out 00:17:25.950 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000601464 s, 6.8 MB/s 00:17:25.950 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:17:25.950 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # size=4096 00:17:25.950 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:17:25.950 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:17:25.950 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@893 -- # return 0 00:17:25.950 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:17:25.950 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:17:25.950 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:17:25.950 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:17:25.950 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=4096 count=7936 oflag=direct 00:17:26.521 7936+0 records in 00:17:26.521 7936+0 records out 00:17:26.521 32505856 bytes (33 MB, 31 MiB) copied, 0.577776 s, 56.3 MB/s 00:17:26.521 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:17:26.521 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:17:26.521 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:17:26.521 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # local nbd_list 00:17:26.521 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@51 -- # local i 00:17:26.521 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:17:26.521 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:17:26.780 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:17:26.780 [2024-11-19 15:23:16.900677] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:26.780 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:17:26.780 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:17:26.780 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:17:26.780 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:17:26.780 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:17:26.780 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@41 -- # break 00:17:26.780 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@45 -- # return 0 00:17:26.780 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:17:26.780 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:26.780 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:26.780 [2024-11-19 15:23:16.916756] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:17:26.780 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:26.780 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:26.781 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:26.781 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:26.781 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:26.781 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:26.781 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:26.781 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:26.781 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:26.781 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:26.781 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:26.781 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:26.781 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:26.781 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:26.781 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:26.781 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:26.781 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:26.781 "name": "raid_bdev1", 00:17:26.781 "uuid": "24259e17-22de-45a2-b4ae-a059493f30d4", 00:17:26.781 "strip_size_kb": 0, 00:17:26.781 "state": "online", 00:17:26.781 "raid_level": "raid1", 00:17:26.781 "superblock": true, 00:17:26.781 "num_base_bdevs": 2, 00:17:26.781 "num_base_bdevs_discovered": 1, 00:17:26.781 "num_base_bdevs_operational": 1, 00:17:26.781 "base_bdevs_list": [ 00:17:26.781 { 00:17:26.781 "name": null, 00:17:26.781 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:26.781 "is_configured": false, 00:17:26.781 "data_offset": 0, 00:17:26.781 "data_size": 7936 00:17:26.781 }, 00:17:26.781 { 00:17:26.781 "name": "BaseBdev2", 00:17:26.781 "uuid": "38aec866-01f4-5eb5-9ba1-80c65063c764", 00:17:26.781 "is_configured": true, 00:17:26.781 "data_offset": 256, 00:17:26.781 "data_size": 7936 00:17:26.781 } 00:17:26.781 ] 00:17:26.781 }' 00:17:26.781 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:26.781 15:23:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:27.041 15:23:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:17:27.041 15:23:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:27.041 15:23:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:27.041 [2024-11-19 15:23:17.324111] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:17:27.041 [2024-11-19 15:23:17.328719] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00019c960 00:17:27.041 15:23:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:27.041 15:23:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@647 -- # sleep 1 00:17:27.041 [2024-11-19 15:23:17.330967] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:17:28.422 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:28.423 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:28.423 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:28.423 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:28.423 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:28.423 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:28.423 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:28.423 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:28.423 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:28.423 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:28.423 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:28.423 "name": "raid_bdev1", 00:17:28.423 "uuid": "24259e17-22de-45a2-b4ae-a059493f30d4", 00:17:28.423 "strip_size_kb": 0, 00:17:28.423 "state": "online", 00:17:28.423 "raid_level": "raid1", 00:17:28.423 "superblock": true, 00:17:28.423 "num_base_bdevs": 2, 00:17:28.423 "num_base_bdevs_discovered": 2, 00:17:28.423 "num_base_bdevs_operational": 2, 00:17:28.423 "process": { 00:17:28.423 "type": "rebuild", 00:17:28.423 "target": "spare", 00:17:28.423 "progress": { 00:17:28.423 "blocks": 2560, 00:17:28.423 "percent": 32 00:17:28.423 } 00:17:28.423 }, 00:17:28.423 "base_bdevs_list": [ 00:17:28.423 { 00:17:28.423 "name": "spare", 00:17:28.423 "uuid": "62e5d54a-3eea-531b-b459-0cfb0f02b978", 00:17:28.423 "is_configured": true, 00:17:28.423 "data_offset": 256, 00:17:28.423 "data_size": 7936 00:17:28.423 }, 00:17:28.423 { 00:17:28.423 "name": "BaseBdev2", 00:17:28.423 "uuid": "38aec866-01f4-5eb5-9ba1-80c65063c764", 00:17:28.423 "is_configured": true, 00:17:28.423 "data_offset": 256, 00:17:28.423 "data_size": 7936 00:17:28.423 } 00:17:28.423 ] 00:17:28.423 }' 00:17:28.423 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:28.423 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:28.423 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:28.423 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:28.423 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:17:28.423 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:28.423 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:28.423 [2024-11-19 15:23:18.479674] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:28.423 [2024-11-19 15:23:18.539502] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:17:28.423 [2024-11-19 15:23:18.539564] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:28.423 [2024-11-19 15:23:18.539600] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:28.423 [2024-11-19 15:23:18.539608] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:17:28.423 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:28.423 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:28.423 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:28.423 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:28.423 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:28.423 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:28.423 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:28.423 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:28.423 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:28.423 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:28.423 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:28.423 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:28.423 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:28.423 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:28.423 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:28.423 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:28.423 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:28.423 "name": "raid_bdev1", 00:17:28.423 "uuid": "24259e17-22de-45a2-b4ae-a059493f30d4", 00:17:28.423 "strip_size_kb": 0, 00:17:28.423 "state": "online", 00:17:28.423 "raid_level": "raid1", 00:17:28.423 "superblock": true, 00:17:28.423 "num_base_bdevs": 2, 00:17:28.423 "num_base_bdevs_discovered": 1, 00:17:28.423 "num_base_bdevs_operational": 1, 00:17:28.423 "base_bdevs_list": [ 00:17:28.423 { 00:17:28.423 "name": null, 00:17:28.423 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:28.423 "is_configured": false, 00:17:28.423 "data_offset": 0, 00:17:28.423 "data_size": 7936 00:17:28.423 }, 00:17:28.423 { 00:17:28.423 "name": "BaseBdev2", 00:17:28.423 "uuid": "38aec866-01f4-5eb5-9ba1-80c65063c764", 00:17:28.423 "is_configured": true, 00:17:28.423 "data_offset": 256, 00:17:28.423 "data_size": 7936 00:17:28.423 } 00:17:28.423 ] 00:17:28.423 }' 00:17:28.423 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:28.423 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:28.683 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:28.683 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:28.683 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:28.683 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:28.683 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:28.683 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:28.683 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:28.683 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:28.683 15:23:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:28.683 15:23:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:28.943 15:23:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:28.943 "name": "raid_bdev1", 00:17:28.943 "uuid": "24259e17-22de-45a2-b4ae-a059493f30d4", 00:17:28.943 "strip_size_kb": 0, 00:17:28.943 "state": "online", 00:17:28.943 "raid_level": "raid1", 00:17:28.943 "superblock": true, 00:17:28.943 "num_base_bdevs": 2, 00:17:28.943 "num_base_bdevs_discovered": 1, 00:17:28.943 "num_base_bdevs_operational": 1, 00:17:28.943 "base_bdevs_list": [ 00:17:28.943 { 00:17:28.943 "name": null, 00:17:28.943 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:28.943 "is_configured": false, 00:17:28.943 "data_offset": 0, 00:17:28.943 "data_size": 7936 00:17:28.943 }, 00:17:28.943 { 00:17:28.943 "name": "BaseBdev2", 00:17:28.943 "uuid": "38aec866-01f4-5eb5-9ba1-80c65063c764", 00:17:28.943 "is_configured": true, 00:17:28.943 "data_offset": 256, 00:17:28.943 "data_size": 7936 00:17:28.943 } 00:17:28.943 ] 00:17:28.943 }' 00:17:28.943 15:23:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:28.943 15:23:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:28.943 15:23:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:28.943 15:23:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:28.943 15:23:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:17:28.943 15:23:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:28.943 15:23:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:28.943 [2024-11-19 15:23:19.140128] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:17:28.943 [2024-11-19 15:23:19.144378] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00019ca30 00:17:28.943 15:23:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:28.943 15:23:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@663 -- # sleep 1 00:17:28.943 [2024-11-19 15:23:19.146570] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:17:29.883 15:23:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:29.883 15:23:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:29.883 15:23:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:29.883 15:23:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:29.883 15:23:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:29.883 15:23:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:29.883 15:23:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:29.883 15:23:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:29.883 15:23:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:29.883 15:23:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:29.883 15:23:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:29.883 "name": "raid_bdev1", 00:17:29.883 "uuid": "24259e17-22de-45a2-b4ae-a059493f30d4", 00:17:29.883 "strip_size_kb": 0, 00:17:29.883 "state": "online", 00:17:29.883 "raid_level": "raid1", 00:17:29.883 "superblock": true, 00:17:29.883 "num_base_bdevs": 2, 00:17:29.883 "num_base_bdevs_discovered": 2, 00:17:29.883 "num_base_bdevs_operational": 2, 00:17:29.883 "process": { 00:17:29.883 "type": "rebuild", 00:17:29.883 "target": "spare", 00:17:29.883 "progress": { 00:17:29.883 "blocks": 2560, 00:17:29.883 "percent": 32 00:17:29.883 } 00:17:29.883 }, 00:17:29.883 "base_bdevs_list": [ 00:17:29.883 { 00:17:29.883 "name": "spare", 00:17:29.883 "uuid": "62e5d54a-3eea-531b-b459-0cfb0f02b978", 00:17:29.883 "is_configured": true, 00:17:29.883 "data_offset": 256, 00:17:29.883 "data_size": 7936 00:17:29.883 }, 00:17:29.883 { 00:17:29.883 "name": "BaseBdev2", 00:17:29.883 "uuid": "38aec866-01f4-5eb5-9ba1-80c65063c764", 00:17:29.883 "is_configured": true, 00:17:29.883 "data_offset": 256, 00:17:29.883 "data_size": 7936 00:17:29.883 } 00:17:29.883 ] 00:17:29.883 }' 00:17:29.883 15:23:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:30.144 15:23:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:30.144 15:23:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:30.144 15:23:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:30.144 15:23:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:17:30.144 15:23:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:17:30.144 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:17:30.144 15:23:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:17:30.144 15:23:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:17:30.144 15:23:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:17:30.144 15:23:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@706 -- # local timeout=598 00:17:30.144 15:23:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:17:30.144 15:23:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:30.144 15:23:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:30.144 15:23:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:30.144 15:23:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:30.144 15:23:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:30.144 15:23:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:30.144 15:23:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:30.144 15:23:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:30.144 15:23:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:30.144 15:23:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:30.144 15:23:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:30.144 "name": "raid_bdev1", 00:17:30.144 "uuid": "24259e17-22de-45a2-b4ae-a059493f30d4", 00:17:30.144 "strip_size_kb": 0, 00:17:30.144 "state": "online", 00:17:30.144 "raid_level": "raid1", 00:17:30.144 "superblock": true, 00:17:30.144 "num_base_bdevs": 2, 00:17:30.144 "num_base_bdevs_discovered": 2, 00:17:30.144 "num_base_bdevs_operational": 2, 00:17:30.144 "process": { 00:17:30.144 "type": "rebuild", 00:17:30.144 "target": "spare", 00:17:30.144 "progress": { 00:17:30.144 "blocks": 2816, 00:17:30.144 "percent": 35 00:17:30.144 } 00:17:30.144 }, 00:17:30.144 "base_bdevs_list": [ 00:17:30.144 { 00:17:30.144 "name": "spare", 00:17:30.144 "uuid": "62e5d54a-3eea-531b-b459-0cfb0f02b978", 00:17:30.144 "is_configured": true, 00:17:30.144 "data_offset": 256, 00:17:30.144 "data_size": 7936 00:17:30.144 }, 00:17:30.144 { 00:17:30.144 "name": "BaseBdev2", 00:17:30.144 "uuid": "38aec866-01f4-5eb5-9ba1-80c65063c764", 00:17:30.144 "is_configured": true, 00:17:30.144 "data_offset": 256, 00:17:30.144 "data_size": 7936 00:17:30.144 } 00:17:30.144 ] 00:17:30.144 }' 00:17:30.144 15:23:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:30.144 15:23:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:30.144 15:23:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:30.144 15:23:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:30.144 15:23:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@711 -- # sleep 1 00:17:31.527 15:23:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:17:31.527 15:23:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:31.527 15:23:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:31.527 15:23:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:31.527 15:23:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:31.527 15:23:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:31.527 15:23:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:31.527 15:23:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:31.527 15:23:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:31.527 15:23:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:31.527 15:23:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:31.527 15:23:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:31.527 "name": "raid_bdev1", 00:17:31.527 "uuid": "24259e17-22de-45a2-b4ae-a059493f30d4", 00:17:31.527 "strip_size_kb": 0, 00:17:31.527 "state": "online", 00:17:31.527 "raid_level": "raid1", 00:17:31.527 "superblock": true, 00:17:31.527 "num_base_bdevs": 2, 00:17:31.527 "num_base_bdevs_discovered": 2, 00:17:31.527 "num_base_bdevs_operational": 2, 00:17:31.527 "process": { 00:17:31.527 "type": "rebuild", 00:17:31.527 "target": "spare", 00:17:31.527 "progress": { 00:17:31.527 "blocks": 5888, 00:17:31.527 "percent": 74 00:17:31.527 } 00:17:31.527 }, 00:17:31.527 "base_bdevs_list": [ 00:17:31.527 { 00:17:31.527 "name": "spare", 00:17:31.527 "uuid": "62e5d54a-3eea-531b-b459-0cfb0f02b978", 00:17:31.527 "is_configured": true, 00:17:31.527 "data_offset": 256, 00:17:31.527 "data_size": 7936 00:17:31.527 }, 00:17:31.527 { 00:17:31.527 "name": "BaseBdev2", 00:17:31.527 "uuid": "38aec866-01f4-5eb5-9ba1-80c65063c764", 00:17:31.527 "is_configured": true, 00:17:31.527 "data_offset": 256, 00:17:31.527 "data_size": 7936 00:17:31.527 } 00:17:31.527 ] 00:17:31.527 }' 00:17:31.527 15:23:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:31.527 15:23:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:31.527 15:23:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:31.527 15:23:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:31.527 15:23:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@711 -- # sleep 1 00:17:32.097 [2024-11-19 15:23:22.260263] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:17:32.097 [2024-11-19 15:23:22.260336] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:17:32.097 [2024-11-19 15:23:22.260438] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:32.358 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:17:32.358 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:32.358 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:32.358 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:32.358 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:32.358 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:32.358 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:32.358 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:32.358 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:32.358 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:32.358 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:32.358 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:32.358 "name": "raid_bdev1", 00:17:32.358 "uuid": "24259e17-22de-45a2-b4ae-a059493f30d4", 00:17:32.358 "strip_size_kb": 0, 00:17:32.358 "state": "online", 00:17:32.358 "raid_level": "raid1", 00:17:32.358 "superblock": true, 00:17:32.358 "num_base_bdevs": 2, 00:17:32.358 "num_base_bdevs_discovered": 2, 00:17:32.358 "num_base_bdevs_operational": 2, 00:17:32.358 "base_bdevs_list": [ 00:17:32.358 { 00:17:32.358 "name": "spare", 00:17:32.358 "uuid": "62e5d54a-3eea-531b-b459-0cfb0f02b978", 00:17:32.358 "is_configured": true, 00:17:32.358 "data_offset": 256, 00:17:32.358 "data_size": 7936 00:17:32.358 }, 00:17:32.358 { 00:17:32.358 "name": "BaseBdev2", 00:17:32.358 "uuid": "38aec866-01f4-5eb5-9ba1-80c65063c764", 00:17:32.358 "is_configured": true, 00:17:32.358 "data_offset": 256, 00:17:32.358 "data_size": 7936 00:17:32.358 } 00:17:32.358 ] 00:17:32.358 }' 00:17:32.358 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:32.619 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:17:32.620 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:32.620 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:17:32.620 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@709 -- # break 00:17:32.620 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:32.620 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:32.620 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:32.620 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:32.620 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:32.620 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:32.620 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:32.620 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:32.620 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:32.620 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:32.620 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:32.620 "name": "raid_bdev1", 00:17:32.620 "uuid": "24259e17-22de-45a2-b4ae-a059493f30d4", 00:17:32.620 "strip_size_kb": 0, 00:17:32.620 "state": "online", 00:17:32.620 "raid_level": "raid1", 00:17:32.620 "superblock": true, 00:17:32.620 "num_base_bdevs": 2, 00:17:32.620 "num_base_bdevs_discovered": 2, 00:17:32.620 "num_base_bdevs_operational": 2, 00:17:32.620 "base_bdevs_list": [ 00:17:32.620 { 00:17:32.620 "name": "spare", 00:17:32.620 "uuid": "62e5d54a-3eea-531b-b459-0cfb0f02b978", 00:17:32.620 "is_configured": true, 00:17:32.620 "data_offset": 256, 00:17:32.620 "data_size": 7936 00:17:32.620 }, 00:17:32.620 { 00:17:32.620 "name": "BaseBdev2", 00:17:32.620 "uuid": "38aec866-01f4-5eb5-9ba1-80c65063c764", 00:17:32.620 "is_configured": true, 00:17:32.620 "data_offset": 256, 00:17:32.620 "data_size": 7936 00:17:32.620 } 00:17:32.620 ] 00:17:32.620 }' 00:17:32.620 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:32.620 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:32.620 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:32.620 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:32.620 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:17:32.620 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:32.620 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:32.620 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:32.620 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:32.620 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:32.620 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:32.620 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:32.620 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:32.620 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:32.620 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:32.620 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:32.620 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:32.620 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:32.620 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:32.620 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:32.620 "name": "raid_bdev1", 00:17:32.620 "uuid": "24259e17-22de-45a2-b4ae-a059493f30d4", 00:17:32.620 "strip_size_kb": 0, 00:17:32.620 "state": "online", 00:17:32.620 "raid_level": "raid1", 00:17:32.620 "superblock": true, 00:17:32.620 "num_base_bdevs": 2, 00:17:32.620 "num_base_bdevs_discovered": 2, 00:17:32.620 "num_base_bdevs_operational": 2, 00:17:32.620 "base_bdevs_list": [ 00:17:32.620 { 00:17:32.620 "name": "spare", 00:17:32.620 "uuid": "62e5d54a-3eea-531b-b459-0cfb0f02b978", 00:17:32.620 "is_configured": true, 00:17:32.620 "data_offset": 256, 00:17:32.620 "data_size": 7936 00:17:32.620 }, 00:17:32.620 { 00:17:32.620 "name": "BaseBdev2", 00:17:32.620 "uuid": "38aec866-01f4-5eb5-9ba1-80c65063c764", 00:17:32.620 "is_configured": true, 00:17:32.620 "data_offset": 256, 00:17:32.620 "data_size": 7936 00:17:32.620 } 00:17:32.620 ] 00:17:32.620 }' 00:17:32.620 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:32.620 15:23:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:33.191 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:17:33.191 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:33.191 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:33.191 [2024-11-19 15:23:23.400957] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:17:33.191 [2024-11-19 15:23:23.400995] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:17:33.191 [2024-11-19 15:23:23.401079] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:33.191 [2024-11-19 15:23:23.401146] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:33.191 [2024-11-19 15:23:23.401160] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:17:33.191 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:33.191 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:33.191 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:33.191 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:33.191 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@720 -- # jq length 00:17:33.191 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:33.191 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:17:33.191 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:17:33.191 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:17:33.191 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:17:33.191 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:17:33.191 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:17:33.191 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # local bdev_list 00:17:33.191 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:17:33.191 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # local nbd_list 00:17:33.191 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@12 -- # local i 00:17:33.191 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:17:33.191 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:17:33.191 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:17:33.452 /dev/nbd0 00:17:33.452 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:17:33.452 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:17:33.452 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:17:33.452 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@873 -- # local i 00:17:33.452 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:17:33.452 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:17:33.452 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:17:33.452 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@877 -- # break 00:17:33.452 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:17:33.452 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:17:33.452 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:17:33.452 1+0 records in 00:17:33.452 1+0 records out 00:17:33.452 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000348549 s, 11.8 MB/s 00:17:33.452 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:17:33.452 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # size=4096 00:17:33.452 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:17:33.452 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:17:33.452 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@893 -- # return 0 00:17:33.452 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:17:33.452 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:17:33.452 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:17:33.713 /dev/nbd1 00:17:33.713 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:17:33.713 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:17:33.713 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:17:33.713 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@873 -- # local i 00:17:33.713 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:17:33.713 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:17:33.713 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:17:33.713 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@877 -- # break 00:17:33.713 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:17:33.713 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:17:33.713 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:17:33.713 1+0 records in 00:17:33.713 1+0 records out 00:17:33.713 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000360901 s, 11.3 MB/s 00:17:33.713 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:17:33.713 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # size=4096 00:17:33.713 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:17:33.713 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:17:33.713 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@893 -- # return 0 00:17:33.713 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:17:33.713 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:17:33.713 15:23:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:17:33.713 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:17:33.713 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:17:33.713 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:17:33.713 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # local nbd_list 00:17:33.713 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@51 -- # local i 00:17:33.713 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:17:33.713 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:17:33.973 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:17:33.973 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:17:33.973 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:17:33.973 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:17:33.973 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:17:33.973 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:17:33.973 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@41 -- # break 00:17:33.973 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@45 -- # return 0 00:17:33.973 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:17:33.973 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:17:34.233 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:17:34.233 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:17:34.233 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:17:34.233 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:17:34.234 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:17:34.234 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:17:34.234 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@41 -- # break 00:17:34.234 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@45 -- # return 0 00:17:34.234 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:17:34.234 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:17:34.234 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:34.234 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:34.234 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:34.234 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:17:34.234 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:34.234 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:34.234 [2024-11-19 15:23:24.480081] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:17:34.234 [2024-11-19 15:23:24.480463] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:34.234 [2024-11-19 15:23:24.480547] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:17:34.234 [2024-11-19 15:23:24.480605] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:34.234 [2024-11-19 15:23:24.482524] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:34.234 [2024-11-19 15:23:24.482621] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:17:34.234 [2024-11-19 15:23:24.482716] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:17:34.234 [2024-11-19 15:23:24.482758] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:17:34.234 [2024-11-19 15:23:24.482855] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:17:34.234 spare 00:17:34.234 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:34.234 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:17:34.234 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:34.234 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:34.494 [2024-11-19 15:23:24.582752] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:17:34.494 [2024-11-19 15:23:24.582781] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:17:34.494 [2024-11-19 15:23:24.582884] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001bb1b0 00:17:34.494 [2024-11-19 15:23:24.582996] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:17:34.494 [2024-11-19 15:23:24.583012] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:17:34.494 [2024-11-19 15:23:24.583101] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:34.494 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:34.494 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:17:34.494 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:34.494 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:34.494 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:34.494 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:34.494 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:34.494 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:34.494 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:34.494 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:34.494 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:34.494 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:34.494 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:34.494 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:34.494 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:34.494 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:34.494 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:34.494 "name": "raid_bdev1", 00:17:34.494 "uuid": "24259e17-22de-45a2-b4ae-a059493f30d4", 00:17:34.494 "strip_size_kb": 0, 00:17:34.494 "state": "online", 00:17:34.494 "raid_level": "raid1", 00:17:34.494 "superblock": true, 00:17:34.494 "num_base_bdevs": 2, 00:17:34.494 "num_base_bdevs_discovered": 2, 00:17:34.494 "num_base_bdevs_operational": 2, 00:17:34.494 "base_bdevs_list": [ 00:17:34.494 { 00:17:34.494 "name": "spare", 00:17:34.494 "uuid": "62e5d54a-3eea-531b-b459-0cfb0f02b978", 00:17:34.494 "is_configured": true, 00:17:34.494 "data_offset": 256, 00:17:34.494 "data_size": 7936 00:17:34.494 }, 00:17:34.494 { 00:17:34.494 "name": "BaseBdev2", 00:17:34.494 "uuid": "38aec866-01f4-5eb5-9ba1-80c65063c764", 00:17:34.494 "is_configured": true, 00:17:34.494 "data_offset": 256, 00:17:34.494 "data_size": 7936 00:17:34.494 } 00:17:34.494 ] 00:17:34.494 }' 00:17:34.494 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:34.494 15:23:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:34.755 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:34.755 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:34.755 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:34.755 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:34.755 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:34.755 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:34.755 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:34.755 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:34.755 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:34.755 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:34.755 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:34.755 "name": "raid_bdev1", 00:17:34.755 "uuid": "24259e17-22de-45a2-b4ae-a059493f30d4", 00:17:34.755 "strip_size_kb": 0, 00:17:34.755 "state": "online", 00:17:34.755 "raid_level": "raid1", 00:17:34.755 "superblock": true, 00:17:34.755 "num_base_bdevs": 2, 00:17:34.755 "num_base_bdevs_discovered": 2, 00:17:34.755 "num_base_bdevs_operational": 2, 00:17:34.755 "base_bdevs_list": [ 00:17:34.755 { 00:17:34.755 "name": "spare", 00:17:34.755 "uuid": "62e5d54a-3eea-531b-b459-0cfb0f02b978", 00:17:34.755 "is_configured": true, 00:17:34.755 "data_offset": 256, 00:17:34.755 "data_size": 7936 00:17:34.755 }, 00:17:34.755 { 00:17:34.755 "name": "BaseBdev2", 00:17:34.755 "uuid": "38aec866-01f4-5eb5-9ba1-80c65063c764", 00:17:34.755 "is_configured": true, 00:17:34.755 "data_offset": 256, 00:17:34.755 "data_size": 7936 00:17:34.755 } 00:17:34.755 ] 00:17:34.755 }' 00:17:34.755 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:35.015 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:35.015 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:35.015 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:35.015 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:17:35.015 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:35.015 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:35.015 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:35.015 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:35.015 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:17:35.015 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:17:35.015 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:35.015 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:35.015 [2024-11-19 15:23:25.215028] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:35.015 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:35.015 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:35.015 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:35.015 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:35.015 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:35.015 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:35.015 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:35.015 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:35.015 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:35.015 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:35.015 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:35.015 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:35.015 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:35.015 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:35.015 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:35.015 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:35.015 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:35.015 "name": "raid_bdev1", 00:17:35.016 "uuid": "24259e17-22de-45a2-b4ae-a059493f30d4", 00:17:35.016 "strip_size_kb": 0, 00:17:35.016 "state": "online", 00:17:35.016 "raid_level": "raid1", 00:17:35.016 "superblock": true, 00:17:35.016 "num_base_bdevs": 2, 00:17:35.016 "num_base_bdevs_discovered": 1, 00:17:35.016 "num_base_bdevs_operational": 1, 00:17:35.016 "base_bdevs_list": [ 00:17:35.016 { 00:17:35.016 "name": null, 00:17:35.016 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:35.016 "is_configured": false, 00:17:35.016 "data_offset": 0, 00:17:35.016 "data_size": 7936 00:17:35.016 }, 00:17:35.016 { 00:17:35.016 "name": "BaseBdev2", 00:17:35.016 "uuid": "38aec866-01f4-5eb5-9ba1-80c65063c764", 00:17:35.016 "is_configured": true, 00:17:35.016 "data_offset": 256, 00:17:35.016 "data_size": 7936 00:17:35.016 } 00:17:35.016 ] 00:17:35.016 }' 00:17:35.016 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:35.016 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:35.585 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:17:35.585 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:35.585 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:35.585 [2024-11-19 15:23:25.694192] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:17:35.585 [2024-11-19 15:23:25.694342] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:17:35.585 [2024-11-19 15:23:25.694365] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:17:35.585 [2024-11-19 15:23:25.694669] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:17:35.585 [2024-11-19 15:23:25.697047] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001bb280 00:17:35.585 [2024-11-19 15:23:25.698929] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:17:35.585 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:35.585 15:23:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@757 -- # sleep 1 00:17:36.525 15:23:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:36.525 15:23:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:36.525 15:23:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:36.525 15:23:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:36.525 15:23:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:36.525 15:23:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:36.525 15:23:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:36.525 15:23:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:36.525 15:23:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:36.525 15:23:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:36.525 15:23:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:36.525 "name": "raid_bdev1", 00:17:36.525 "uuid": "24259e17-22de-45a2-b4ae-a059493f30d4", 00:17:36.525 "strip_size_kb": 0, 00:17:36.525 "state": "online", 00:17:36.525 "raid_level": "raid1", 00:17:36.525 "superblock": true, 00:17:36.525 "num_base_bdevs": 2, 00:17:36.525 "num_base_bdevs_discovered": 2, 00:17:36.525 "num_base_bdevs_operational": 2, 00:17:36.525 "process": { 00:17:36.525 "type": "rebuild", 00:17:36.525 "target": "spare", 00:17:36.525 "progress": { 00:17:36.525 "blocks": 2560, 00:17:36.525 "percent": 32 00:17:36.525 } 00:17:36.525 }, 00:17:36.525 "base_bdevs_list": [ 00:17:36.525 { 00:17:36.525 "name": "spare", 00:17:36.525 "uuid": "62e5d54a-3eea-531b-b459-0cfb0f02b978", 00:17:36.525 "is_configured": true, 00:17:36.525 "data_offset": 256, 00:17:36.525 "data_size": 7936 00:17:36.525 }, 00:17:36.525 { 00:17:36.525 "name": "BaseBdev2", 00:17:36.525 "uuid": "38aec866-01f4-5eb5-9ba1-80c65063c764", 00:17:36.525 "is_configured": true, 00:17:36.525 "data_offset": 256, 00:17:36.525 "data_size": 7936 00:17:36.525 } 00:17:36.525 ] 00:17:36.525 }' 00:17:36.525 15:23:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:36.525 15:23:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:36.525 15:23:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:36.525 15:23:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:36.525 15:23:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:17:36.525 15:23:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:36.525 15:23:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:36.525 [2024-11-19 15:23:26.853785] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:36.785 [2024-11-19 15:23:26.903526] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:17:36.785 [2024-11-19 15:23:26.903892] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:36.785 [2024-11-19 15:23:26.903924] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:36.785 [2024-11-19 15:23:26.903934] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:17:36.785 15:23:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:36.785 15:23:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:36.785 15:23:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:36.785 15:23:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:36.785 15:23:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:36.785 15:23:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:36.785 15:23:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:36.785 15:23:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:36.785 15:23:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:36.785 15:23:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:36.785 15:23:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:36.785 15:23:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:36.785 15:23:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:36.785 15:23:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:36.785 15:23:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:36.785 15:23:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:36.785 15:23:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:36.785 "name": "raid_bdev1", 00:17:36.785 "uuid": "24259e17-22de-45a2-b4ae-a059493f30d4", 00:17:36.785 "strip_size_kb": 0, 00:17:36.785 "state": "online", 00:17:36.785 "raid_level": "raid1", 00:17:36.785 "superblock": true, 00:17:36.785 "num_base_bdevs": 2, 00:17:36.785 "num_base_bdevs_discovered": 1, 00:17:36.785 "num_base_bdevs_operational": 1, 00:17:36.785 "base_bdevs_list": [ 00:17:36.785 { 00:17:36.785 "name": null, 00:17:36.785 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:36.785 "is_configured": false, 00:17:36.785 "data_offset": 0, 00:17:36.785 "data_size": 7936 00:17:36.785 }, 00:17:36.785 { 00:17:36.785 "name": "BaseBdev2", 00:17:36.785 "uuid": "38aec866-01f4-5eb5-9ba1-80c65063c764", 00:17:36.785 "is_configured": true, 00:17:36.785 "data_offset": 256, 00:17:36.785 "data_size": 7936 00:17:36.785 } 00:17:36.785 ] 00:17:36.785 }' 00:17:36.785 15:23:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:36.785 15:23:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:37.045 15:23:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:17:37.045 15:23:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:37.045 15:23:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:37.045 [2024-11-19 15:23:27.286034] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:17:37.045 [2024-11-19 15:23:27.286273] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:37.045 [2024-11-19 15:23:27.286347] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:17:37.045 [2024-11-19 15:23:27.286402] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:37.045 [2024-11-19 15:23:27.286666] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:37.045 [2024-11-19 15:23:27.286773] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:17:37.045 [2024-11-19 15:23:27.286892] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:17:37.045 [2024-11-19 15:23:27.286934] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:17:37.045 [2024-11-19 15:23:27.287004] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:17:37.045 [2024-11-19 15:23:27.287123] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:17:37.045 [2024-11-19 15:23:27.289463] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001bb350 00:17:37.045 [2024-11-19 15:23:27.291441] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:17:37.045 spare 00:17:37.045 15:23:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:37.045 15:23:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@764 -- # sleep 1 00:17:37.985 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:37.985 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:37.985 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:37.985 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:37.985 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:37.986 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:37.986 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:37.986 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:37.986 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:37.986 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:38.245 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:38.245 "name": "raid_bdev1", 00:17:38.245 "uuid": "24259e17-22de-45a2-b4ae-a059493f30d4", 00:17:38.245 "strip_size_kb": 0, 00:17:38.245 "state": "online", 00:17:38.245 "raid_level": "raid1", 00:17:38.245 "superblock": true, 00:17:38.245 "num_base_bdevs": 2, 00:17:38.245 "num_base_bdevs_discovered": 2, 00:17:38.245 "num_base_bdevs_operational": 2, 00:17:38.245 "process": { 00:17:38.245 "type": "rebuild", 00:17:38.245 "target": "spare", 00:17:38.245 "progress": { 00:17:38.245 "blocks": 2560, 00:17:38.245 "percent": 32 00:17:38.245 } 00:17:38.245 }, 00:17:38.245 "base_bdevs_list": [ 00:17:38.245 { 00:17:38.245 "name": "spare", 00:17:38.245 "uuid": "62e5d54a-3eea-531b-b459-0cfb0f02b978", 00:17:38.245 "is_configured": true, 00:17:38.245 "data_offset": 256, 00:17:38.245 "data_size": 7936 00:17:38.245 }, 00:17:38.245 { 00:17:38.245 "name": "BaseBdev2", 00:17:38.245 "uuid": "38aec866-01f4-5eb5-9ba1-80c65063c764", 00:17:38.245 "is_configured": true, 00:17:38.245 "data_offset": 256, 00:17:38.245 "data_size": 7936 00:17:38.245 } 00:17:38.245 ] 00:17:38.245 }' 00:17:38.246 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:38.246 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:38.246 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:38.246 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:38.246 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:17:38.246 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:38.246 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:38.246 [2024-11-19 15:23:28.454266] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:38.246 [2024-11-19 15:23:28.495492] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:17:38.246 [2024-11-19 15:23:28.495886] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:38.246 [2024-11-19 15:23:28.495942] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:38.246 [2024-11-19 15:23:28.495977] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:17:38.246 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:38.246 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:38.246 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:38.246 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:38.246 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:38.246 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:38.246 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:38.246 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:38.246 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:38.246 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:38.246 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:38.246 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:38.246 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:38.246 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:38.246 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:38.246 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:38.246 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:38.246 "name": "raid_bdev1", 00:17:38.246 "uuid": "24259e17-22de-45a2-b4ae-a059493f30d4", 00:17:38.246 "strip_size_kb": 0, 00:17:38.246 "state": "online", 00:17:38.246 "raid_level": "raid1", 00:17:38.246 "superblock": true, 00:17:38.246 "num_base_bdevs": 2, 00:17:38.246 "num_base_bdevs_discovered": 1, 00:17:38.246 "num_base_bdevs_operational": 1, 00:17:38.246 "base_bdevs_list": [ 00:17:38.246 { 00:17:38.246 "name": null, 00:17:38.246 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:38.246 "is_configured": false, 00:17:38.246 "data_offset": 0, 00:17:38.246 "data_size": 7936 00:17:38.246 }, 00:17:38.246 { 00:17:38.246 "name": "BaseBdev2", 00:17:38.246 "uuid": "38aec866-01f4-5eb5-9ba1-80c65063c764", 00:17:38.246 "is_configured": true, 00:17:38.246 "data_offset": 256, 00:17:38.246 "data_size": 7936 00:17:38.246 } 00:17:38.246 ] 00:17:38.246 }' 00:17:38.246 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:38.246 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:38.816 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:38.816 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:38.816 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:38.816 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:38.816 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:38.816 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:38.816 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:38.816 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:38.816 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:38.816 15:23:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:38.816 15:23:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:38.816 "name": "raid_bdev1", 00:17:38.816 "uuid": "24259e17-22de-45a2-b4ae-a059493f30d4", 00:17:38.816 "strip_size_kb": 0, 00:17:38.816 "state": "online", 00:17:38.816 "raid_level": "raid1", 00:17:38.816 "superblock": true, 00:17:38.816 "num_base_bdevs": 2, 00:17:38.816 "num_base_bdevs_discovered": 1, 00:17:38.816 "num_base_bdevs_operational": 1, 00:17:38.816 "base_bdevs_list": [ 00:17:38.816 { 00:17:38.816 "name": null, 00:17:38.816 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:38.816 "is_configured": false, 00:17:38.816 "data_offset": 0, 00:17:38.816 "data_size": 7936 00:17:38.816 }, 00:17:38.816 { 00:17:38.816 "name": "BaseBdev2", 00:17:38.816 "uuid": "38aec866-01f4-5eb5-9ba1-80c65063c764", 00:17:38.816 "is_configured": true, 00:17:38.816 "data_offset": 256, 00:17:38.816 "data_size": 7936 00:17:38.816 } 00:17:38.816 ] 00:17:38.816 }' 00:17:38.816 15:23:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:38.816 15:23:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:38.816 15:23:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:38.816 15:23:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:38.816 15:23:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:17:38.816 15:23:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:38.816 15:23:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:38.816 15:23:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:38.816 15:23:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:17:38.816 15:23:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:38.816 15:23:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:38.816 [2024-11-19 15:23:29.125516] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:17:38.816 [2024-11-19 15:23:29.125816] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:38.816 [2024-11-19 15:23:29.125895] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:17:38.816 [2024-11-19 15:23:29.126002] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:38.816 [2024-11-19 15:23:29.126253] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:38.816 [2024-11-19 15:23:29.126365] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:17:38.816 [2024-11-19 15:23:29.126480] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:17:38.816 [2024-11-19 15:23:29.126542] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:17:38.816 [2024-11-19 15:23:29.126578] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:17:38.816 [2024-11-19 15:23:29.126615] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:17:38.816 BaseBdev1 00:17:38.816 15:23:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:38.816 15:23:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@775 -- # sleep 1 00:17:40.197 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:40.197 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:40.197 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:40.197 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:40.197 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:40.197 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:40.197 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:40.197 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:40.197 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:40.197 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:40.197 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:40.197 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:40.197 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:40.197 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:40.197 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:40.197 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:40.197 "name": "raid_bdev1", 00:17:40.197 "uuid": "24259e17-22de-45a2-b4ae-a059493f30d4", 00:17:40.197 "strip_size_kb": 0, 00:17:40.197 "state": "online", 00:17:40.197 "raid_level": "raid1", 00:17:40.197 "superblock": true, 00:17:40.197 "num_base_bdevs": 2, 00:17:40.197 "num_base_bdevs_discovered": 1, 00:17:40.197 "num_base_bdevs_operational": 1, 00:17:40.197 "base_bdevs_list": [ 00:17:40.197 { 00:17:40.197 "name": null, 00:17:40.197 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:40.197 "is_configured": false, 00:17:40.197 "data_offset": 0, 00:17:40.197 "data_size": 7936 00:17:40.197 }, 00:17:40.197 { 00:17:40.197 "name": "BaseBdev2", 00:17:40.197 "uuid": "38aec866-01f4-5eb5-9ba1-80c65063c764", 00:17:40.197 "is_configured": true, 00:17:40.197 "data_offset": 256, 00:17:40.197 "data_size": 7936 00:17:40.197 } 00:17:40.197 ] 00:17:40.197 }' 00:17:40.197 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:40.197 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:40.457 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:40.457 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:40.457 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:40.457 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:40.457 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:40.457 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:40.457 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:40.457 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:40.457 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:40.457 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:40.457 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:40.457 "name": "raid_bdev1", 00:17:40.457 "uuid": "24259e17-22de-45a2-b4ae-a059493f30d4", 00:17:40.457 "strip_size_kb": 0, 00:17:40.457 "state": "online", 00:17:40.457 "raid_level": "raid1", 00:17:40.457 "superblock": true, 00:17:40.457 "num_base_bdevs": 2, 00:17:40.457 "num_base_bdevs_discovered": 1, 00:17:40.457 "num_base_bdevs_operational": 1, 00:17:40.457 "base_bdevs_list": [ 00:17:40.457 { 00:17:40.457 "name": null, 00:17:40.457 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:40.457 "is_configured": false, 00:17:40.457 "data_offset": 0, 00:17:40.457 "data_size": 7936 00:17:40.457 }, 00:17:40.457 { 00:17:40.457 "name": "BaseBdev2", 00:17:40.457 "uuid": "38aec866-01f4-5eb5-9ba1-80c65063c764", 00:17:40.457 "is_configured": true, 00:17:40.457 "data_offset": 256, 00:17:40.457 "data_size": 7936 00:17:40.457 } 00:17:40.457 ] 00:17:40.457 }' 00:17:40.457 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:40.457 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:40.457 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:40.457 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:40.457 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:17:40.457 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@652 -- # local es=0 00:17:40.457 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:17:40.457 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:17:40.457 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:17:40.457 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:17:40.457 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:17:40.457 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:17:40.457 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:40.457 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:40.457 [2024-11-19 15:23:30.743003] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:17:40.457 [2024-11-19 15:23:30.743154] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:17:40.457 [2024-11-19 15:23:30.743223] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:17:40.457 request: 00:17:40.457 { 00:17:40.457 "base_bdev": "BaseBdev1", 00:17:40.457 "raid_bdev": "raid_bdev1", 00:17:40.457 "method": "bdev_raid_add_base_bdev", 00:17:40.457 "req_id": 1 00:17:40.457 } 00:17:40.457 Got JSON-RPC error response 00:17:40.457 response: 00:17:40.457 { 00:17:40.457 "code": -22, 00:17:40.457 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:17:40.457 } 00:17:40.457 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:17:40.457 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@655 -- # es=1 00:17:40.457 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:17:40.457 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:17:40.457 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:17:40.457 15:23:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@779 -- # sleep 1 00:17:41.839 15:23:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:41.839 15:23:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:41.839 15:23:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:41.839 15:23:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:41.839 15:23:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:41.839 15:23:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:41.839 15:23:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:41.839 15:23:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:41.839 15:23:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:41.839 15:23:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:41.839 15:23:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:41.839 15:23:31 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:41.839 15:23:31 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:41.839 15:23:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:41.839 15:23:31 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:41.839 15:23:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:41.839 "name": "raid_bdev1", 00:17:41.839 "uuid": "24259e17-22de-45a2-b4ae-a059493f30d4", 00:17:41.839 "strip_size_kb": 0, 00:17:41.839 "state": "online", 00:17:41.839 "raid_level": "raid1", 00:17:41.839 "superblock": true, 00:17:41.839 "num_base_bdevs": 2, 00:17:41.839 "num_base_bdevs_discovered": 1, 00:17:41.839 "num_base_bdevs_operational": 1, 00:17:41.839 "base_bdevs_list": [ 00:17:41.839 { 00:17:41.839 "name": null, 00:17:41.839 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:41.839 "is_configured": false, 00:17:41.839 "data_offset": 0, 00:17:41.839 "data_size": 7936 00:17:41.839 }, 00:17:41.839 { 00:17:41.839 "name": "BaseBdev2", 00:17:41.839 "uuid": "38aec866-01f4-5eb5-9ba1-80c65063c764", 00:17:41.839 "is_configured": true, 00:17:41.839 "data_offset": 256, 00:17:41.839 "data_size": 7936 00:17:41.839 } 00:17:41.839 ] 00:17:41.839 }' 00:17:41.839 15:23:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:41.839 15:23:31 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:42.099 15:23:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:42.099 15:23:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:42.099 15:23:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:42.099 15:23:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:42.099 15:23:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:42.099 15:23:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:42.099 15:23:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:42.099 15:23:32 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:42.099 15:23:32 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:42.099 15:23:32 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:42.099 15:23:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:42.099 "name": "raid_bdev1", 00:17:42.099 "uuid": "24259e17-22de-45a2-b4ae-a059493f30d4", 00:17:42.099 "strip_size_kb": 0, 00:17:42.099 "state": "online", 00:17:42.099 "raid_level": "raid1", 00:17:42.099 "superblock": true, 00:17:42.099 "num_base_bdevs": 2, 00:17:42.099 "num_base_bdevs_discovered": 1, 00:17:42.099 "num_base_bdevs_operational": 1, 00:17:42.099 "base_bdevs_list": [ 00:17:42.099 { 00:17:42.099 "name": null, 00:17:42.099 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:42.099 "is_configured": false, 00:17:42.099 "data_offset": 0, 00:17:42.099 "data_size": 7936 00:17:42.099 }, 00:17:42.099 { 00:17:42.099 "name": "BaseBdev2", 00:17:42.099 "uuid": "38aec866-01f4-5eb5-9ba1-80c65063c764", 00:17:42.099 "is_configured": true, 00:17:42.099 "data_offset": 256, 00:17:42.099 "data_size": 7936 00:17:42.099 } 00:17:42.099 ] 00:17:42.099 }' 00:17:42.099 15:23:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:42.099 15:23:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:42.099 15:23:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:42.099 15:23:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:42.099 15:23:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@784 -- # killprocess 98104 00:17:42.099 15:23:32 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@954 -- # '[' -z 98104 ']' 00:17:42.099 15:23:32 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@958 -- # kill -0 98104 00:17:42.099 15:23:32 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@959 -- # uname 00:17:42.099 15:23:32 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:17:42.099 15:23:32 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 98104 00:17:42.099 15:23:32 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:17:42.099 killing process with pid 98104 00:17:42.099 Received shutdown signal, test time was about 60.000000 seconds 00:17:42.099 00:17:42.099 Latency(us) 00:17:42.099 [2024-11-19T15:23:32.438Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:17:42.099 [2024-11-19T15:23:32.438Z] =================================================================================================================== 00:17:42.099 [2024-11-19T15:23:32.438Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:17:42.099 15:23:32 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:17:42.099 15:23:32 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@972 -- # echo 'killing process with pid 98104' 00:17:42.099 15:23:32 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@973 -- # kill 98104 00:17:42.099 [2024-11-19 15:23:32.417878] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:17:42.099 [2024-11-19 15:23:32.417979] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:42.099 [2024-11-19 15:23:32.418022] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:42.099 [2024-11-19 15:23:32.418031] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:17:42.099 15:23:32 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@978 -- # wait 98104 00:17:42.359 [2024-11-19 15:23:32.451556] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:17:42.359 ************************************ 00:17:42.359 END TEST raid_rebuild_test_sb_md_separate 00:17:42.359 ************************************ 00:17:42.359 15:23:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@786 -- # return 0 00:17:42.359 00:17:42.359 real 0m18.407s 00:17:42.359 user 0m24.440s 00:17:42.359 sys 0m2.679s 00:17:42.359 15:23:32 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@1130 -- # xtrace_disable 00:17:42.359 15:23:32 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:17:42.619 15:23:32 bdev_raid -- bdev/bdev_raid.sh@1010 -- # base_malloc_params='-m 32 -i' 00:17:42.619 15:23:32 bdev_raid -- bdev/bdev_raid.sh@1011 -- # run_test raid_state_function_test_sb_md_interleaved raid_state_function_test raid1 2 true 00:17:42.619 15:23:32 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:17:42.619 15:23:32 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:17:42.619 15:23:32 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:17:42.619 ************************************ 00:17:42.619 START TEST raid_state_function_test_sb_md_interleaved 00:17:42.619 ************************************ 00:17:42.619 15:23:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 2 true 00:17:42.619 15:23:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:17:42.619 15:23:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:17:42.620 15:23:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:17:42.620 15:23:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:17:42.620 15:23:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:17:42.620 15:23:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:17:42.620 15:23:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:17:42.620 15:23:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:17:42.620 15:23:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:17:42.620 15:23:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:17:42.620 15:23:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:17:42.620 15:23:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:17:42.620 15:23:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:17:42.620 15:23:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:17:42.620 15:23:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:17:42.620 15:23:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@211 -- # local strip_size 00:17:42.620 15:23:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:17:42.620 15:23:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:17:42.620 15:23:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:17:42.620 15:23:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:17:42.620 15:23:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:17:42.620 15:23:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:17:42.620 15:23:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@229 -- # raid_pid=98793 00:17:42.620 15:23:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:17:42.620 15:23:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 98793' 00:17:42.620 Process raid pid: 98793 00:17:42.620 15:23:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@231 -- # waitforlisten 98793 00:17:42.620 15:23:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@835 -- # '[' -z 98793 ']' 00:17:42.620 15:23:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:17:42.620 15:23:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@840 -- # local max_retries=100 00:17:42.620 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:17:42.620 15:23:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:17:42.620 15:23:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@844 -- # xtrace_disable 00:17:42.620 15:23:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:42.620 [2024-11-19 15:23:32.823348] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:17:42.620 [2024-11-19 15:23:32.823470] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:17:42.882 [2024-11-19 15:23:32.979414] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:42.882 [2024-11-19 15:23:33.004711] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:42.882 [2024-11-19 15:23:33.047352] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:17:42.882 [2024-11-19 15:23:33.047391] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:17:43.460 15:23:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:17:43.460 15:23:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@868 -- # return 0 00:17:43.460 15:23:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:17:43.460 15:23:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:43.460 15:23:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:43.460 [2024-11-19 15:23:33.648827] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:17:43.460 [2024-11-19 15:23:33.648877] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:17:43.460 [2024-11-19 15:23:33.648886] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:17:43.460 [2024-11-19 15:23:33.648897] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:17:43.460 15:23:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:43.460 15:23:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:17:43.460 15:23:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:17:43.460 15:23:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:17:43.460 15:23:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:43.460 15:23:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:43.460 15:23:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:43.460 15:23:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:43.460 15:23:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:43.460 15:23:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:43.460 15:23:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:43.460 15:23:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:43.460 15:23:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:17:43.460 15:23:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:43.460 15:23:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:43.460 15:23:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:43.460 15:23:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:43.460 "name": "Existed_Raid", 00:17:43.460 "uuid": "491e3249-51aa-4849-b736-52cd2c67c56d", 00:17:43.460 "strip_size_kb": 0, 00:17:43.460 "state": "configuring", 00:17:43.460 "raid_level": "raid1", 00:17:43.460 "superblock": true, 00:17:43.460 "num_base_bdevs": 2, 00:17:43.460 "num_base_bdevs_discovered": 0, 00:17:43.460 "num_base_bdevs_operational": 2, 00:17:43.460 "base_bdevs_list": [ 00:17:43.460 { 00:17:43.460 "name": "BaseBdev1", 00:17:43.460 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:43.460 "is_configured": false, 00:17:43.460 "data_offset": 0, 00:17:43.460 "data_size": 0 00:17:43.460 }, 00:17:43.460 { 00:17:43.460 "name": "BaseBdev2", 00:17:43.460 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:43.460 "is_configured": false, 00:17:43.460 "data_offset": 0, 00:17:43.460 "data_size": 0 00:17:43.460 } 00:17:43.460 ] 00:17:43.460 }' 00:17:43.460 15:23:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:43.460 15:23:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:44.031 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:17:44.031 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:44.031 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:44.031 [2024-11-19 15:23:34.092064] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:17:44.031 [2024-11-19 15:23:34.092183] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:17:44.031 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:44.031 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:17:44.031 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:44.031 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:44.031 [2024-11-19 15:23:34.104031] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:17:44.031 [2024-11-19 15:23:34.104126] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:17:44.031 [2024-11-19 15:23:34.104152] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:17:44.031 [2024-11-19 15:23:34.104185] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:17:44.031 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:44.031 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev1 00:17:44.031 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:44.031 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:44.031 [2024-11-19 15:23:34.124826] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:17:44.031 BaseBdev1 00:17:44.031 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:44.031 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:17:44.031 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:17:44.031 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:17:44.031 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@905 -- # local i 00:17:44.031 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:17:44.031 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:17:44.031 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:17:44.031 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:44.031 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:44.031 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:44.031 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:17:44.031 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:44.031 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:44.031 [ 00:17:44.031 { 00:17:44.031 "name": "BaseBdev1", 00:17:44.031 "aliases": [ 00:17:44.031 "0a75f607-961f-4a79-8465-c24cdfb8ce8c" 00:17:44.031 ], 00:17:44.031 "product_name": "Malloc disk", 00:17:44.031 "block_size": 4128, 00:17:44.031 "num_blocks": 8192, 00:17:44.031 "uuid": "0a75f607-961f-4a79-8465-c24cdfb8ce8c", 00:17:44.031 "md_size": 32, 00:17:44.031 "md_interleave": true, 00:17:44.031 "dif_type": 0, 00:17:44.031 "assigned_rate_limits": { 00:17:44.031 "rw_ios_per_sec": 0, 00:17:44.031 "rw_mbytes_per_sec": 0, 00:17:44.031 "r_mbytes_per_sec": 0, 00:17:44.031 "w_mbytes_per_sec": 0 00:17:44.031 }, 00:17:44.031 "claimed": true, 00:17:44.031 "claim_type": "exclusive_write", 00:17:44.031 "zoned": false, 00:17:44.031 "supported_io_types": { 00:17:44.031 "read": true, 00:17:44.031 "write": true, 00:17:44.031 "unmap": true, 00:17:44.031 "flush": true, 00:17:44.031 "reset": true, 00:17:44.031 "nvme_admin": false, 00:17:44.031 "nvme_io": false, 00:17:44.031 "nvme_io_md": false, 00:17:44.031 "write_zeroes": true, 00:17:44.031 "zcopy": true, 00:17:44.031 "get_zone_info": false, 00:17:44.031 "zone_management": false, 00:17:44.031 "zone_append": false, 00:17:44.031 "compare": false, 00:17:44.031 "compare_and_write": false, 00:17:44.031 "abort": true, 00:17:44.031 "seek_hole": false, 00:17:44.031 "seek_data": false, 00:17:44.031 "copy": true, 00:17:44.031 "nvme_iov_md": false 00:17:44.031 }, 00:17:44.031 "memory_domains": [ 00:17:44.031 { 00:17:44.032 "dma_device_id": "system", 00:17:44.032 "dma_device_type": 1 00:17:44.032 }, 00:17:44.032 { 00:17:44.032 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:44.032 "dma_device_type": 2 00:17:44.032 } 00:17:44.032 ], 00:17:44.032 "driver_specific": {} 00:17:44.032 } 00:17:44.032 ] 00:17:44.032 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:44.032 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@911 -- # return 0 00:17:44.032 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:17:44.032 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:17:44.032 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:17:44.032 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:44.032 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:44.032 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:44.032 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:44.032 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:44.032 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:44.032 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:44.032 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:17:44.032 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:44.032 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:44.032 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:44.032 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:44.032 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:44.032 "name": "Existed_Raid", 00:17:44.032 "uuid": "6bd25285-4657-42cf-854c-2f11a5941ae0", 00:17:44.032 "strip_size_kb": 0, 00:17:44.032 "state": "configuring", 00:17:44.032 "raid_level": "raid1", 00:17:44.032 "superblock": true, 00:17:44.032 "num_base_bdevs": 2, 00:17:44.032 "num_base_bdevs_discovered": 1, 00:17:44.032 "num_base_bdevs_operational": 2, 00:17:44.032 "base_bdevs_list": [ 00:17:44.032 { 00:17:44.032 "name": "BaseBdev1", 00:17:44.032 "uuid": "0a75f607-961f-4a79-8465-c24cdfb8ce8c", 00:17:44.032 "is_configured": true, 00:17:44.032 "data_offset": 256, 00:17:44.032 "data_size": 7936 00:17:44.032 }, 00:17:44.032 { 00:17:44.032 "name": "BaseBdev2", 00:17:44.032 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:44.032 "is_configured": false, 00:17:44.032 "data_offset": 0, 00:17:44.032 "data_size": 0 00:17:44.032 } 00:17:44.032 ] 00:17:44.032 }' 00:17:44.032 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:44.032 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:44.601 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:17:44.601 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:44.601 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:44.601 [2024-11-19 15:23:34.639974] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:17:44.601 [2024-11-19 15:23:34.640028] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:17:44.601 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:44.601 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:17:44.601 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:44.601 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:44.601 [2024-11-19 15:23:34.652028] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:17:44.601 [2024-11-19 15:23:34.653859] bdev.c:8278:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:17:44.601 [2024-11-19 15:23:34.653898] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:17:44.601 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:44.601 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:17:44.601 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:17:44.601 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:17:44.601 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:17:44.601 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:17:44.601 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:44.601 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:44.601 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:44.601 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:44.601 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:44.601 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:44.601 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:44.601 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:17:44.601 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:44.601 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:44.601 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:44.601 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:44.601 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:44.601 "name": "Existed_Raid", 00:17:44.601 "uuid": "7b2d0fff-21fa-4e40-a115-c105dcb3367c", 00:17:44.601 "strip_size_kb": 0, 00:17:44.601 "state": "configuring", 00:17:44.601 "raid_level": "raid1", 00:17:44.601 "superblock": true, 00:17:44.601 "num_base_bdevs": 2, 00:17:44.601 "num_base_bdevs_discovered": 1, 00:17:44.601 "num_base_bdevs_operational": 2, 00:17:44.601 "base_bdevs_list": [ 00:17:44.601 { 00:17:44.601 "name": "BaseBdev1", 00:17:44.601 "uuid": "0a75f607-961f-4a79-8465-c24cdfb8ce8c", 00:17:44.602 "is_configured": true, 00:17:44.602 "data_offset": 256, 00:17:44.602 "data_size": 7936 00:17:44.602 }, 00:17:44.602 { 00:17:44.602 "name": "BaseBdev2", 00:17:44.602 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:44.602 "is_configured": false, 00:17:44.602 "data_offset": 0, 00:17:44.602 "data_size": 0 00:17:44.602 } 00:17:44.602 ] 00:17:44.602 }' 00:17:44.602 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:44.602 15:23:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:44.862 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev2 00:17:44.862 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:44.862 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:44.862 [2024-11-19 15:23:35.094186] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:17:44.862 [2024-11-19 15:23:35.094409] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:17:44.862 [2024-11-19 15:23:35.094466] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:17:44.862 [2024-11-19 15:23:35.094581] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:17:44.862 [2024-11-19 15:23:35.094700] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:17:44.862 [2024-11-19 15:23:35.094751] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:17:44.862 [2024-11-19 15:23:35.094845] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:44.862 BaseBdev2 00:17:44.862 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:44.862 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:17:44.862 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:17:44.862 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:17:44.862 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@905 -- # local i 00:17:44.862 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:17:44.862 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:17:44.862 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:17:44.862 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:44.862 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:44.862 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:44.862 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:17:44.862 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:44.862 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:44.862 [ 00:17:44.862 { 00:17:44.862 "name": "BaseBdev2", 00:17:44.862 "aliases": [ 00:17:44.862 "8cbcf795-6853-4970-962a-36b98b575596" 00:17:44.862 ], 00:17:44.862 "product_name": "Malloc disk", 00:17:44.862 "block_size": 4128, 00:17:44.862 "num_blocks": 8192, 00:17:44.862 "uuid": "8cbcf795-6853-4970-962a-36b98b575596", 00:17:44.862 "md_size": 32, 00:17:44.862 "md_interleave": true, 00:17:44.862 "dif_type": 0, 00:17:44.862 "assigned_rate_limits": { 00:17:44.862 "rw_ios_per_sec": 0, 00:17:44.862 "rw_mbytes_per_sec": 0, 00:17:44.862 "r_mbytes_per_sec": 0, 00:17:44.862 "w_mbytes_per_sec": 0 00:17:44.862 }, 00:17:44.862 "claimed": true, 00:17:44.862 "claim_type": "exclusive_write", 00:17:44.862 "zoned": false, 00:17:44.862 "supported_io_types": { 00:17:44.862 "read": true, 00:17:44.862 "write": true, 00:17:44.862 "unmap": true, 00:17:44.862 "flush": true, 00:17:44.862 "reset": true, 00:17:44.862 "nvme_admin": false, 00:17:44.862 "nvme_io": false, 00:17:44.862 "nvme_io_md": false, 00:17:44.862 "write_zeroes": true, 00:17:44.862 "zcopy": true, 00:17:44.862 "get_zone_info": false, 00:17:44.862 "zone_management": false, 00:17:44.862 "zone_append": false, 00:17:44.862 "compare": false, 00:17:44.862 "compare_and_write": false, 00:17:44.862 "abort": true, 00:17:44.862 "seek_hole": false, 00:17:44.862 "seek_data": false, 00:17:44.862 "copy": true, 00:17:44.862 "nvme_iov_md": false 00:17:44.862 }, 00:17:44.862 "memory_domains": [ 00:17:44.862 { 00:17:44.862 "dma_device_id": "system", 00:17:44.862 "dma_device_type": 1 00:17:44.862 }, 00:17:44.862 { 00:17:44.862 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:44.862 "dma_device_type": 2 00:17:44.862 } 00:17:44.862 ], 00:17:44.862 "driver_specific": {} 00:17:44.862 } 00:17:44.862 ] 00:17:44.862 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:44.862 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@911 -- # return 0 00:17:44.862 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:17:44.862 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:17:44.862 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:17:44.862 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:17:44.862 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:44.862 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:44.862 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:44.862 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:44.862 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:44.862 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:44.862 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:44.862 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:44.862 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:44.862 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:44.862 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:44.862 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:17:44.862 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:44.862 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:44.862 "name": "Existed_Raid", 00:17:44.862 "uuid": "7b2d0fff-21fa-4e40-a115-c105dcb3367c", 00:17:44.862 "strip_size_kb": 0, 00:17:44.862 "state": "online", 00:17:44.862 "raid_level": "raid1", 00:17:44.862 "superblock": true, 00:17:44.862 "num_base_bdevs": 2, 00:17:44.862 "num_base_bdevs_discovered": 2, 00:17:44.862 "num_base_bdevs_operational": 2, 00:17:44.862 "base_bdevs_list": [ 00:17:44.862 { 00:17:44.862 "name": "BaseBdev1", 00:17:44.862 "uuid": "0a75f607-961f-4a79-8465-c24cdfb8ce8c", 00:17:44.862 "is_configured": true, 00:17:44.862 "data_offset": 256, 00:17:44.862 "data_size": 7936 00:17:44.862 }, 00:17:44.862 { 00:17:44.862 "name": "BaseBdev2", 00:17:44.862 "uuid": "8cbcf795-6853-4970-962a-36b98b575596", 00:17:44.862 "is_configured": true, 00:17:44.862 "data_offset": 256, 00:17:44.862 "data_size": 7936 00:17:44.862 } 00:17:44.862 ] 00:17:44.862 }' 00:17:44.862 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:44.862 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:45.433 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:17:45.433 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:17:45.433 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:17:45.433 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:17:45.433 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@184 -- # local name 00:17:45.433 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:17:45.433 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:17:45.433 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:17:45.433 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:45.433 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:45.433 [2024-11-19 15:23:35.585631] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:17:45.433 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:45.433 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:17:45.433 "name": "Existed_Raid", 00:17:45.433 "aliases": [ 00:17:45.433 "7b2d0fff-21fa-4e40-a115-c105dcb3367c" 00:17:45.433 ], 00:17:45.433 "product_name": "Raid Volume", 00:17:45.433 "block_size": 4128, 00:17:45.433 "num_blocks": 7936, 00:17:45.433 "uuid": "7b2d0fff-21fa-4e40-a115-c105dcb3367c", 00:17:45.433 "md_size": 32, 00:17:45.433 "md_interleave": true, 00:17:45.433 "dif_type": 0, 00:17:45.433 "assigned_rate_limits": { 00:17:45.433 "rw_ios_per_sec": 0, 00:17:45.433 "rw_mbytes_per_sec": 0, 00:17:45.433 "r_mbytes_per_sec": 0, 00:17:45.433 "w_mbytes_per_sec": 0 00:17:45.433 }, 00:17:45.433 "claimed": false, 00:17:45.433 "zoned": false, 00:17:45.433 "supported_io_types": { 00:17:45.433 "read": true, 00:17:45.433 "write": true, 00:17:45.433 "unmap": false, 00:17:45.433 "flush": false, 00:17:45.433 "reset": true, 00:17:45.433 "nvme_admin": false, 00:17:45.433 "nvme_io": false, 00:17:45.433 "nvme_io_md": false, 00:17:45.433 "write_zeroes": true, 00:17:45.433 "zcopy": false, 00:17:45.433 "get_zone_info": false, 00:17:45.433 "zone_management": false, 00:17:45.433 "zone_append": false, 00:17:45.433 "compare": false, 00:17:45.433 "compare_and_write": false, 00:17:45.433 "abort": false, 00:17:45.433 "seek_hole": false, 00:17:45.433 "seek_data": false, 00:17:45.433 "copy": false, 00:17:45.433 "nvme_iov_md": false 00:17:45.433 }, 00:17:45.433 "memory_domains": [ 00:17:45.433 { 00:17:45.433 "dma_device_id": "system", 00:17:45.433 "dma_device_type": 1 00:17:45.433 }, 00:17:45.433 { 00:17:45.433 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:45.433 "dma_device_type": 2 00:17:45.433 }, 00:17:45.433 { 00:17:45.433 "dma_device_id": "system", 00:17:45.433 "dma_device_type": 1 00:17:45.433 }, 00:17:45.433 { 00:17:45.433 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:45.433 "dma_device_type": 2 00:17:45.433 } 00:17:45.433 ], 00:17:45.433 "driver_specific": { 00:17:45.433 "raid": { 00:17:45.433 "uuid": "7b2d0fff-21fa-4e40-a115-c105dcb3367c", 00:17:45.433 "strip_size_kb": 0, 00:17:45.433 "state": "online", 00:17:45.433 "raid_level": "raid1", 00:17:45.433 "superblock": true, 00:17:45.433 "num_base_bdevs": 2, 00:17:45.433 "num_base_bdevs_discovered": 2, 00:17:45.433 "num_base_bdevs_operational": 2, 00:17:45.433 "base_bdevs_list": [ 00:17:45.433 { 00:17:45.433 "name": "BaseBdev1", 00:17:45.433 "uuid": "0a75f607-961f-4a79-8465-c24cdfb8ce8c", 00:17:45.433 "is_configured": true, 00:17:45.433 "data_offset": 256, 00:17:45.433 "data_size": 7936 00:17:45.433 }, 00:17:45.433 { 00:17:45.433 "name": "BaseBdev2", 00:17:45.433 "uuid": "8cbcf795-6853-4970-962a-36b98b575596", 00:17:45.433 "is_configured": true, 00:17:45.433 "data_offset": 256, 00:17:45.433 "data_size": 7936 00:17:45.433 } 00:17:45.433 ] 00:17:45.433 } 00:17:45.433 } 00:17:45.433 }' 00:17:45.433 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:17:45.433 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:17:45.433 BaseBdev2' 00:17:45.433 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:17:45.433 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4128 32 true 0' 00:17:45.433 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:17:45.433 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:17:45.433 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:45.433 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:45.433 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:17:45.433 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:45.433 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:17:45.433 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:17:45.433 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:17:45.433 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:17:45.433 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:17:45.433 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:45.433 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:45.693 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:45.693 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:17:45.693 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:17:45.693 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:17:45.693 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:45.693 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:45.693 [2024-11-19 15:23:35.809071] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:17:45.693 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:45.693 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@260 -- # local expected_state 00:17:45.693 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:17:45.693 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@198 -- # case $1 in 00:17:45.693 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@199 -- # return 0 00:17:45.693 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:17:45.693 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:17:45.693 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:17:45.693 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:45.693 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:45.693 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:45.693 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:45.693 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:45.693 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:45.693 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:45.693 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:45.693 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:45.693 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:17:45.693 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:45.693 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:45.693 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:45.693 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:45.693 "name": "Existed_Raid", 00:17:45.693 "uuid": "7b2d0fff-21fa-4e40-a115-c105dcb3367c", 00:17:45.693 "strip_size_kb": 0, 00:17:45.693 "state": "online", 00:17:45.693 "raid_level": "raid1", 00:17:45.693 "superblock": true, 00:17:45.693 "num_base_bdevs": 2, 00:17:45.693 "num_base_bdevs_discovered": 1, 00:17:45.693 "num_base_bdevs_operational": 1, 00:17:45.693 "base_bdevs_list": [ 00:17:45.693 { 00:17:45.693 "name": null, 00:17:45.693 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:45.693 "is_configured": false, 00:17:45.693 "data_offset": 0, 00:17:45.693 "data_size": 7936 00:17:45.693 }, 00:17:45.693 { 00:17:45.693 "name": "BaseBdev2", 00:17:45.693 "uuid": "8cbcf795-6853-4970-962a-36b98b575596", 00:17:45.693 "is_configured": true, 00:17:45.693 "data_offset": 256, 00:17:45.693 "data_size": 7936 00:17:45.693 } 00:17:45.693 ] 00:17:45.693 }' 00:17:45.693 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:45.693 15:23:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:45.953 15:23:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:17:45.953 15:23:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:17:45.953 15:23:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:45.953 15:23:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:45.953 15:23:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:17:45.953 15:23:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:45.953 15:23:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:45.953 15:23:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:17:45.953 15:23:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:17:45.953 15:23:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:17:45.953 15:23:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:45.953 15:23:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:45.953 [2024-11-19 15:23:36.287719] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:17:45.953 [2024-11-19 15:23:36.287885] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:17:46.213 [2024-11-19 15:23:36.299811] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:46.213 [2024-11-19 15:23:36.299991] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:46.213 [2024-11-19 15:23:36.300010] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:17:46.213 15:23:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:46.213 15:23:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:17:46.213 15:23:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:17:46.213 15:23:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:46.213 15:23:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:46.213 15:23:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:17:46.213 15:23:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:46.213 15:23:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:46.213 15:23:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:17:46.213 15:23:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:17:46.213 15:23:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:17:46.213 15:23:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@326 -- # killprocess 98793 00:17:46.213 15:23:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@954 -- # '[' -z 98793 ']' 00:17:46.213 15:23:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@958 -- # kill -0 98793 00:17:46.213 15:23:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@959 -- # uname 00:17:46.213 15:23:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:17:46.213 15:23:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 98793 00:17:46.213 15:23:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:17:46.213 15:23:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:17:46.213 killing process with pid 98793 00:17:46.213 15:23:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@972 -- # echo 'killing process with pid 98793' 00:17:46.213 15:23:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@973 -- # kill 98793 00:17:46.213 [2024-11-19 15:23:36.397145] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:17:46.213 15:23:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@978 -- # wait 98793 00:17:46.213 [2024-11-19 15:23:36.398097] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:17:46.473 15:23:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@328 -- # return 0 00:17:46.473 00:17:46.473 real 0m3.883s 00:17:46.473 user 0m6.134s 00:17:46.473 sys 0m0.832s 00:17:46.473 15:23:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@1130 -- # xtrace_disable 00:17:46.473 15:23:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:46.473 ************************************ 00:17:46.473 END TEST raid_state_function_test_sb_md_interleaved 00:17:46.473 ************************************ 00:17:46.473 15:23:36 bdev_raid -- bdev/bdev_raid.sh@1012 -- # run_test raid_superblock_test_md_interleaved raid_superblock_test raid1 2 00:17:46.473 15:23:36 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:17:46.473 15:23:36 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:17:46.473 15:23:36 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:17:46.473 ************************************ 00:17:46.473 START TEST raid_superblock_test_md_interleaved 00:17:46.473 ************************************ 00:17:46.473 15:23:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@1129 -- # raid_superblock_test raid1 2 00:17:46.473 15:23:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:17:46.473 15:23:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:17:46.473 15:23:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:17:46.474 15:23:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:17:46.474 15:23:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:17:46.474 15:23:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:17:46.474 15:23:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:17:46.474 15:23:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:17:46.474 15:23:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:17:46.474 15:23:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@399 -- # local strip_size 00:17:46.474 15:23:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:17:46.474 15:23:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:17:46.474 15:23:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:17:46.474 15:23:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:17:46.474 15:23:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:17:46.474 15:23:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@412 -- # raid_pid=99031 00:17:46.474 15:23:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:17:46.474 15:23:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@413 -- # waitforlisten 99031 00:17:46.474 15:23:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@835 -- # '[' -z 99031 ']' 00:17:46.474 15:23:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:17:46.474 15:23:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@840 -- # local max_retries=100 00:17:46.474 15:23:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:17:46.474 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:17:46.474 15:23:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@844 -- # xtrace_disable 00:17:46.474 15:23:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:46.474 [2024-11-19 15:23:36.793526] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:17:46.474 [2024-11-19 15:23:36.793660] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid99031 ] 00:17:46.734 [2024-11-19 15:23:36.947707] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:46.734 [2024-11-19 15:23:36.972114] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:46.734 [2024-11-19 15:23:37.014737] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:17:46.734 [2024-11-19 15:23:37.014864] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:17:47.304 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:17:47.304 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@868 -- # return 0 00:17:47.304 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:17:47.304 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:17:47.304 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:17:47.304 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:17:47.304 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:17:47.304 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:17:47.304 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:17:47.304 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:17:47.304 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b malloc1 00:17:47.304 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:47.304 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:47.304 malloc1 00:17:47.304 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:47.304 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:17:47.304 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:47.304 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:47.304 [2024-11-19 15:23:37.629261] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:17:47.304 [2024-11-19 15:23:37.629410] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:47.304 [2024-11-19 15:23:37.629451] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:17:47.304 [2024-11-19 15:23:37.629462] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:47.304 [2024-11-19 15:23:37.631303] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:47.305 [2024-11-19 15:23:37.631343] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:17:47.305 pt1 00:17:47.305 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:47.305 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:17:47.305 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:17:47.305 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:17:47.305 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:17:47.305 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:17:47.305 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:17:47.305 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:17:47.305 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:17:47.305 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b malloc2 00:17:47.305 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:47.305 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:47.564 malloc2 00:17:47.564 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:47.564 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:17:47.564 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:47.564 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:47.564 [2024-11-19 15:23:37.657904] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:17:47.564 [2024-11-19 15:23:37.658024] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:47.564 [2024-11-19 15:23:37.658074] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:17:47.564 [2024-11-19 15:23:37.658103] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:47.564 [2024-11-19 15:23:37.659945] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:47.564 [2024-11-19 15:23:37.660024] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:17:47.564 pt2 00:17:47.564 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:47.564 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:17:47.564 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:17:47.564 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:17:47.564 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:47.564 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:47.564 [2024-11-19 15:23:37.669912] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:17:47.564 [2024-11-19 15:23:37.671755] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:17:47.564 [2024-11-19 15:23:37.671952] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:17:47.564 [2024-11-19 15:23:37.672021] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:17:47.564 [2024-11-19 15:23:37.672122] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:17:47.564 [2024-11-19 15:23:37.672239] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:17:47.564 [2024-11-19 15:23:37.672280] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:17:47.564 [2024-11-19 15:23:37.672392] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:47.564 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:47.564 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:17:47.564 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:47.564 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:47.564 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:47.564 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:47.564 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:47.564 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:47.564 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:47.564 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:47.564 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:47.564 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:47.564 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:47.564 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:47.564 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:47.564 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:47.564 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:47.564 "name": "raid_bdev1", 00:17:47.564 "uuid": "f4aab9be-5639-4d45-a9be-7b917d7dc3dc", 00:17:47.564 "strip_size_kb": 0, 00:17:47.564 "state": "online", 00:17:47.564 "raid_level": "raid1", 00:17:47.564 "superblock": true, 00:17:47.564 "num_base_bdevs": 2, 00:17:47.564 "num_base_bdevs_discovered": 2, 00:17:47.564 "num_base_bdevs_operational": 2, 00:17:47.564 "base_bdevs_list": [ 00:17:47.564 { 00:17:47.565 "name": "pt1", 00:17:47.565 "uuid": "00000000-0000-0000-0000-000000000001", 00:17:47.565 "is_configured": true, 00:17:47.565 "data_offset": 256, 00:17:47.565 "data_size": 7936 00:17:47.565 }, 00:17:47.565 { 00:17:47.565 "name": "pt2", 00:17:47.565 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:47.565 "is_configured": true, 00:17:47.565 "data_offset": 256, 00:17:47.565 "data_size": 7936 00:17:47.565 } 00:17:47.565 ] 00:17:47.565 }' 00:17:47.565 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:47.565 15:23:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:47.825 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:17:47.825 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:17:47.825 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:17:47.825 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:17:47.825 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@184 -- # local name 00:17:47.825 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:17:47.825 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:17:47.825 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:47.825 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:47.825 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:17:47.825 [2024-11-19 15:23:38.105379] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:17:47.825 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:47.825 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:17:47.825 "name": "raid_bdev1", 00:17:47.825 "aliases": [ 00:17:47.825 "f4aab9be-5639-4d45-a9be-7b917d7dc3dc" 00:17:47.825 ], 00:17:47.825 "product_name": "Raid Volume", 00:17:47.825 "block_size": 4128, 00:17:47.825 "num_blocks": 7936, 00:17:47.825 "uuid": "f4aab9be-5639-4d45-a9be-7b917d7dc3dc", 00:17:47.825 "md_size": 32, 00:17:47.825 "md_interleave": true, 00:17:47.825 "dif_type": 0, 00:17:47.825 "assigned_rate_limits": { 00:17:47.825 "rw_ios_per_sec": 0, 00:17:47.825 "rw_mbytes_per_sec": 0, 00:17:47.825 "r_mbytes_per_sec": 0, 00:17:47.825 "w_mbytes_per_sec": 0 00:17:47.825 }, 00:17:47.825 "claimed": false, 00:17:47.825 "zoned": false, 00:17:47.825 "supported_io_types": { 00:17:47.825 "read": true, 00:17:47.825 "write": true, 00:17:47.825 "unmap": false, 00:17:47.825 "flush": false, 00:17:47.825 "reset": true, 00:17:47.825 "nvme_admin": false, 00:17:47.825 "nvme_io": false, 00:17:47.825 "nvme_io_md": false, 00:17:47.825 "write_zeroes": true, 00:17:47.825 "zcopy": false, 00:17:47.825 "get_zone_info": false, 00:17:47.825 "zone_management": false, 00:17:47.825 "zone_append": false, 00:17:47.825 "compare": false, 00:17:47.825 "compare_and_write": false, 00:17:47.825 "abort": false, 00:17:47.825 "seek_hole": false, 00:17:47.825 "seek_data": false, 00:17:47.825 "copy": false, 00:17:47.825 "nvme_iov_md": false 00:17:47.825 }, 00:17:47.825 "memory_domains": [ 00:17:47.825 { 00:17:47.825 "dma_device_id": "system", 00:17:47.825 "dma_device_type": 1 00:17:47.825 }, 00:17:47.825 { 00:17:47.825 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:47.825 "dma_device_type": 2 00:17:47.825 }, 00:17:47.825 { 00:17:47.825 "dma_device_id": "system", 00:17:47.825 "dma_device_type": 1 00:17:47.825 }, 00:17:47.825 { 00:17:47.825 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:47.825 "dma_device_type": 2 00:17:47.825 } 00:17:47.825 ], 00:17:47.825 "driver_specific": { 00:17:47.825 "raid": { 00:17:47.825 "uuid": "f4aab9be-5639-4d45-a9be-7b917d7dc3dc", 00:17:47.825 "strip_size_kb": 0, 00:17:47.825 "state": "online", 00:17:47.825 "raid_level": "raid1", 00:17:47.825 "superblock": true, 00:17:47.825 "num_base_bdevs": 2, 00:17:47.825 "num_base_bdevs_discovered": 2, 00:17:47.825 "num_base_bdevs_operational": 2, 00:17:47.825 "base_bdevs_list": [ 00:17:47.825 { 00:17:47.825 "name": "pt1", 00:17:47.825 "uuid": "00000000-0000-0000-0000-000000000001", 00:17:47.825 "is_configured": true, 00:17:47.825 "data_offset": 256, 00:17:47.825 "data_size": 7936 00:17:47.825 }, 00:17:47.825 { 00:17:47.825 "name": "pt2", 00:17:47.825 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:47.825 "is_configured": true, 00:17:47.825 "data_offset": 256, 00:17:47.825 "data_size": 7936 00:17:47.825 } 00:17:47.825 ] 00:17:47.825 } 00:17:47.825 } 00:17:47.825 }' 00:17:47.825 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:17:48.085 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:17:48.085 pt2' 00:17:48.085 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:17:48.085 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4128 32 true 0' 00:17:48.085 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:17:48.085 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:17:48.085 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:17:48.085 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:48.085 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:48.085 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:48.085 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:17:48.085 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:17:48.085 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:17:48.085 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:17:48.085 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:17:48.085 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:48.085 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:48.085 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:48.085 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:17:48.085 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:17:48.085 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:17:48.085 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:17:48.085 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:48.085 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:48.085 [2024-11-19 15:23:38.356909] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:17:48.085 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:48.085 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=f4aab9be-5639-4d45-a9be-7b917d7dc3dc 00:17:48.085 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@436 -- # '[' -z f4aab9be-5639-4d45-a9be-7b917d7dc3dc ']' 00:17:48.085 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:17:48.085 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:48.085 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:48.085 [2024-11-19 15:23:38.404635] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:17:48.085 [2024-11-19 15:23:38.404706] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:17:48.085 [2024-11-19 15:23:38.404809] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:48.085 [2024-11-19 15:23:38.404884] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:48.085 [2024-11-19 15:23:38.404920] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:17:48.085 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:48.085 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:48.085 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:17:48.085 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:48.085 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:48.346 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:48.346 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:17:48.346 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:17:48.346 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:17:48.346 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:17:48.346 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:48.346 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:48.346 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:48.346 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:17:48.346 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:17:48.346 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:48.346 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:48.346 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:48.346 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:17:48.346 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:48.346 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:48.346 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:17:48.346 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:48.346 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:17:48.346 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:17:48.346 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@652 -- # local es=0 00:17:48.346 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:17:48.346 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:17:48.346 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:17:48.346 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:17:48.346 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:17:48.346 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:17:48.346 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:48.346 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:48.346 [2024-11-19 15:23:38.540401] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:17:48.346 [2024-11-19 15:23:38.542286] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:17:48.346 [2024-11-19 15:23:38.542385] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:17:48.346 [2024-11-19 15:23:38.542488] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:17:48.346 [2024-11-19 15:23:38.542531] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:17:48.346 [2024-11-19 15:23:38.542552] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:17:48.346 request: 00:17:48.346 { 00:17:48.346 "name": "raid_bdev1", 00:17:48.346 "raid_level": "raid1", 00:17:48.346 "base_bdevs": [ 00:17:48.346 "malloc1", 00:17:48.346 "malloc2" 00:17:48.346 ], 00:17:48.346 "superblock": false, 00:17:48.346 "method": "bdev_raid_create", 00:17:48.346 "req_id": 1 00:17:48.346 } 00:17:48.346 Got JSON-RPC error response 00:17:48.346 response: 00:17:48.346 { 00:17:48.346 "code": -17, 00:17:48.346 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:17:48.346 } 00:17:48.346 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:17:48.346 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@655 -- # es=1 00:17:48.346 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:17:48.346 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:17:48.346 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:17:48.346 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:48.346 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:48.346 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:48.346 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:17:48.346 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:48.346 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:17:48.346 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:17:48.346 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:17:48.346 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:48.346 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:48.346 [2024-11-19 15:23:38.604259] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:17:48.346 [2024-11-19 15:23:38.604310] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:48.346 [2024-11-19 15:23:38.604342] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:17:48.346 [2024-11-19 15:23:38.604350] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:48.346 [2024-11-19 15:23:38.606188] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:48.346 [2024-11-19 15:23:38.606271] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:17:48.347 [2024-11-19 15:23:38.606315] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:17:48.347 [2024-11-19 15:23:38.606355] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:17:48.347 pt1 00:17:48.347 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:48.347 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:17:48.347 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:48.347 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:17:48.347 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:48.347 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:48.347 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:48.347 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:48.347 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:48.347 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:48.347 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:48.347 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:48.347 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:48.347 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:48.347 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:48.347 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:48.347 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:48.347 "name": "raid_bdev1", 00:17:48.347 "uuid": "f4aab9be-5639-4d45-a9be-7b917d7dc3dc", 00:17:48.347 "strip_size_kb": 0, 00:17:48.347 "state": "configuring", 00:17:48.347 "raid_level": "raid1", 00:17:48.347 "superblock": true, 00:17:48.347 "num_base_bdevs": 2, 00:17:48.347 "num_base_bdevs_discovered": 1, 00:17:48.347 "num_base_bdevs_operational": 2, 00:17:48.347 "base_bdevs_list": [ 00:17:48.347 { 00:17:48.347 "name": "pt1", 00:17:48.347 "uuid": "00000000-0000-0000-0000-000000000001", 00:17:48.347 "is_configured": true, 00:17:48.347 "data_offset": 256, 00:17:48.347 "data_size": 7936 00:17:48.347 }, 00:17:48.347 { 00:17:48.347 "name": null, 00:17:48.347 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:48.347 "is_configured": false, 00:17:48.347 "data_offset": 256, 00:17:48.347 "data_size": 7936 00:17:48.347 } 00:17:48.347 ] 00:17:48.347 }' 00:17:48.347 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:48.347 15:23:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:48.918 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:17:48.918 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:17:48.918 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:17:48.918 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:17:48.918 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:48.918 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:48.918 [2024-11-19 15:23:39.007630] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:17:48.918 [2024-11-19 15:23:39.007733] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:48.918 [2024-11-19 15:23:39.007766] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:17:48.918 [2024-11-19 15:23:39.007792] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:48.918 [2024-11-19 15:23:39.007927] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:48.918 [2024-11-19 15:23:39.007985] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:17:48.918 [2024-11-19 15:23:39.008046] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:17:48.918 [2024-11-19 15:23:39.008095] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:17:48.918 [2024-11-19 15:23:39.008189] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:17:48.918 [2024-11-19 15:23:39.008223] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:17:48.918 [2024-11-19 15:23:39.008304] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:17:48.918 [2024-11-19 15:23:39.008384] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:17:48.918 [2024-11-19 15:23:39.008424] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:17:48.918 [2024-11-19 15:23:39.008501] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:48.918 pt2 00:17:48.918 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:48.918 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:17:48.918 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:17:48.918 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:17:48.918 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:48.918 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:48.918 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:48.918 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:48.918 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:48.918 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:48.918 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:48.918 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:48.918 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:48.918 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:48.918 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:48.918 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:48.918 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:48.918 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:48.918 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:48.918 "name": "raid_bdev1", 00:17:48.918 "uuid": "f4aab9be-5639-4d45-a9be-7b917d7dc3dc", 00:17:48.918 "strip_size_kb": 0, 00:17:48.918 "state": "online", 00:17:48.918 "raid_level": "raid1", 00:17:48.918 "superblock": true, 00:17:48.918 "num_base_bdevs": 2, 00:17:48.918 "num_base_bdevs_discovered": 2, 00:17:48.918 "num_base_bdevs_operational": 2, 00:17:48.918 "base_bdevs_list": [ 00:17:48.918 { 00:17:48.918 "name": "pt1", 00:17:48.918 "uuid": "00000000-0000-0000-0000-000000000001", 00:17:48.918 "is_configured": true, 00:17:48.918 "data_offset": 256, 00:17:48.918 "data_size": 7936 00:17:48.918 }, 00:17:48.918 { 00:17:48.918 "name": "pt2", 00:17:48.918 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:48.918 "is_configured": true, 00:17:48.918 "data_offset": 256, 00:17:48.918 "data_size": 7936 00:17:48.918 } 00:17:48.918 ] 00:17:48.918 }' 00:17:48.918 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:48.918 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:49.177 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:17:49.177 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:17:49.177 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:17:49.177 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:17:49.177 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@184 -- # local name 00:17:49.177 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:17:49.177 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:17:49.177 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:17:49.177 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:49.177 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:49.177 [2024-11-19 15:23:39.467079] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:17:49.177 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:49.177 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:17:49.177 "name": "raid_bdev1", 00:17:49.177 "aliases": [ 00:17:49.177 "f4aab9be-5639-4d45-a9be-7b917d7dc3dc" 00:17:49.177 ], 00:17:49.177 "product_name": "Raid Volume", 00:17:49.177 "block_size": 4128, 00:17:49.177 "num_blocks": 7936, 00:17:49.177 "uuid": "f4aab9be-5639-4d45-a9be-7b917d7dc3dc", 00:17:49.177 "md_size": 32, 00:17:49.177 "md_interleave": true, 00:17:49.177 "dif_type": 0, 00:17:49.177 "assigned_rate_limits": { 00:17:49.177 "rw_ios_per_sec": 0, 00:17:49.177 "rw_mbytes_per_sec": 0, 00:17:49.177 "r_mbytes_per_sec": 0, 00:17:49.177 "w_mbytes_per_sec": 0 00:17:49.177 }, 00:17:49.177 "claimed": false, 00:17:49.177 "zoned": false, 00:17:49.177 "supported_io_types": { 00:17:49.177 "read": true, 00:17:49.177 "write": true, 00:17:49.177 "unmap": false, 00:17:49.177 "flush": false, 00:17:49.177 "reset": true, 00:17:49.177 "nvme_admin": false, 00:17:49.177 "nvme_io": false, 00:17:49.177 "nvme_io_md": false, 00:17:49.177 "write_zeroes": true, 00:17:49.177 "zcopy": false, 00:17:49.177 "get_zone_info": false, 00:17:49.177 "zone_management": false, 00:17:49.177 "zone_append": false, 00:17:49.177 "compare": false, 00:17:49.177 "compare_and_write": false, 00:17:49.177 "abort": false, 00:17:49.177 "seek_hole": false, 00:17:49.177 "seek_data": false, 00:17:49.177 "copy": false, 00:17:49.177 "nvme_iov_md": false 00:17:49.177 }, 00:17:49.177 "memory_domains": [ 00:17:49.177 { 00:17:49.177 "dma_device_id": "system", 00:17:49.177 "dma_device_type": 1 00:17:49.177 }, 00:17:49.177 { 00:17:49.177 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:49.177 "dma_device_type": 2 00:17:49.177 }, 00:17:49.177 { 00:17:49.177 "dma_device_id": "system", 00:17:49.177 "dma_device_type": 1 00:17:49.177 }, 00:17:49.177 { 00:17:49.177 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:49.177 "dma_device_type": 2 00:17:49.177 } 00:17:49.177 ], 00:17:49.177 "driver_specific": { 00:17:49.177 "raid": { 00:17:49.177 "uuid": "f4aab9be-5639-4d45-a9be-7b917d7dc3dc", 00:17:49.177 "strip_size_kb": 0, 00:17:49.177 "state": "online", 00:17:49.177 "raid_level": "raid1", 00:17:49.177 "superblock": true, 00:17:49.177 "num_base_bdevs": 2, 00:17:49.177 "num_base_bdevs_discovered": 2, 00:17:49.177 "num_base_bdevs_operational": 2, 00:17:49.177 "base_bdevs_list": [ 00:17:49.177 { 00:17:49.177 "name": "pt1", 00:17:49.177 "uuid": "00000000-0000-0000-0000-000000000001", 00:17:49.177 "is_configured": true, 00:17:49.177 "data_offset": 256, 00:17:49.177 "data_size": 7936 00:17:49.177 }, 00:17:49.177 { 00:17:49.177 "name": "pt2", 00:17:49.177 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:49.177 "is_configured": true, 00:17:49.177 "data_offset": 256, 00:17:49.177 "data_size": 7936 00:17:49.177 } 00:17:49.177 ] 00:17:49.177 } 00:17:49.177 } 00:17:49.177 }' 00:17:49.177 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:17:49.436 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:17:49.436 pt2' 00:17:49.436 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:17:49.436 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4128 32 true 0' 00:17:49.436 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:17:49.436 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:17:49.436 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:17:49.436 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:49.436 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:49.436 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:49.436 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:17:49.436 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:17:49.436 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:17:49.436 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:17:49.436 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:17:49.436 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:49.436 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:49.436 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:49.436 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:17:49.436 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:17:49.436 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:17:49.436 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:17:49.436 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:49.436 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:49.436 [2024-11-19 15:23:39.702670] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:17:49.436 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:49.436 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@487 -- # '[' f4aab9be-5639-4d45-a9be-7b917d7dc3dc '!=' f4aab9be-5639-4d45-a9be-7b917d7dc3dc ']' 00:17:49.436 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:17:49.436 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@198 -- # case $1 in 00:17:49.436 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@199 -- # return 0 00:17:49.436 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:17:49.436 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:49.436 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:49.436 [2024-11-19 15:23:39.750389] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:17:49.436 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:49.436 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:49.436 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:49.436 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:49.436 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:49.436 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:49.436 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:49.436 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:49.436 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:49.436 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:49.436 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:49.437 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:49.437 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:49.437 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:49.437 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:49.697 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:49.697 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:49.697 "name": "raid_bdev1", 00:17:49.697 "uuid": "f4aab9be-5639-4d45-a9be-7b917d7dc3dc", 00:17:49.697 "strip_size_kb": 0, 00:17:49.697 "state": "online", 00:17:49.697 "raid_level": "raid1", 00:17:49.697 "superblock": true, 00:17:49.697 "num_base_bdevs": 2, 00:17:49.697 "num_base_bdevs_discovered": 1, 00:17:49.697 "num_base_bdevs_operational": 1, 00:17:49.697 "base_bdevs_list": [ 00:17:49.697 { 00:17:49.697 "name": null, 00:17:49.697 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:49.697 "is_configured": false, 00:17:49.697 "data_offset": 0, 00:17:49.697 "data_size": 7936 00:17:49.697 }, 00:17:49.697 { 00:17:49.697 "name": "pt2", 00:17:49.697 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:49.697 "is_configured": true, 00:17:49.697 "data_offset": 256, 00:17:49.697 "data_size": 7936 00:17:49.697 } 00:17:49.697 ] 00:17:49.697 }' 00:17:49.697 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:49.697 15:23:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:49.958 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:17:49.958 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:49.958 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:49.958 [2024-11-19 15:23:40.205586] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:17:49.958 [2024-11-19 15:23:40.205657] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:17:49.958 [2024-11-19 15:23:40.205721] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:49.958 [2024-11-19 15:23:40.205771] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:49.958 [2024-11-19 15:23:40.205817] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:17:49.958 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:49.958 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:49.958 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:17:49.958 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:49.958 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:49.958 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:49.958 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:17:49.958 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:17:49.958 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:17:49.958 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:17:49.958 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:17:49.958 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:49.958 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:49.958 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:49.958 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:17:49.958 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:17:49.958 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:17:49.958 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:17:49.958 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@519 -- # i=1 00:17:49.958 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:17:49.958 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:49.958 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:49.958 [2024-11-19 15:23:40.277454] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:17:49.958 [2024-11-19 15:23:40.277547] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:49.958 [2024-11-19 15:23:40.277597] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:17:49.958 [2024-11-19 15:23:40.277623] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:49.958 [2024-11-19 15:23:40.279403] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:49.958 [2024-11-19 15:23:40.279481] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:17:49.958 [2024-11-19 15:23:40.279540] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:17:49.958 [2024-11-19 15:23:40.279582] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:17:49.958 [2024-11-19 15:23:40.279662] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:17:49.958 [2024-11-19 15:23:40.279695] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:17:49.958 [2024-11-19 15:23:40.279774] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:17:49.958 [2024-11-19 15:23:40.279887] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:17:49.958 [2024-11-19 15:23:40.279922] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:17:49.958 [2024-11-19 15:23:40.280014] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:49.958 pt2 00:17:49.958 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:49.958 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:49.958 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:49.958 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:49.958 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:49.958 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:49.958 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:49.958 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:49.958 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:49.958 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:49.958 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:49.958 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:49.958 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:49.958 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:49.958 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:50.219 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:50.219 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:50.219 "name": "raid_bdev1", 00:17:50.220 "uuid": "f4aab9be-5639-4d45-a9be-7b917d7dc3dc", 00:17:50.220 "strip_size_kb": 0, 00:17:50.220 "state": "online", 00:17:50.220 "raid_level": "raid1", 00:17:50.220 "superblock": true, 00:17:50.220 "num_base_bdevs": 2, 00:17:50.220 "num_base_bdevs_discovered": 1, 00:17:50.220 "num_base_bdevs_operational": 1, 00:17:50.220 "base_bdevs_list": [ 00:17:50.220 { 00:17:50.220 "name": null, 00:17:50.220 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:50.220 "is_configured": false, 00:17:50.220 "data_offset": 256, 00:17:50.220 "data_size": 7936 00:17:50.220 }, 00:17:50.220 { 00:17:50.220 "name": "pt2", 00:17:50.220 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:50.220 "is_configured": true, 00:17:50.220 "data_offset": 256, 00:17:50.220 "data_size": 7936 00:17:50.220 } 00:17:50.220 ] 00:17:50.220 }' 00:17:50.220 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:50.220 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:50.480 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:17:50.480 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:50.480 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:50.480 [2024-11-19 15:23:40.696707] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:17:50.480 [2024-11-19 15:23:40.696776] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:17:50.480 [2024-11-19 15:23:40.696838] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:50.480 [2024-11-19 15:23:40.696873] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:50.480 [2024-11-19 15:23:40.696885] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:17:50.480 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:50.480 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:50.480 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:50.480 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:50.480 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:17:50.480 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:50.480 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:17:50.480 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:17:50.480 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:17:50.480 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:17:50.481 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:50.481 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:50.481 [2024-11-19 15:23:40.744640] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:17:50.481 [2024-11-19 15:23:40.744696] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:50.481 [2024-11-19 15:23:40.744712] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008d80 00:17:50.481 [2024-11-19 15:23:40.744725] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:50.481 [2024-11-19 15:23:40.746578] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:50.481 [2024-11-19 15:23:40.746663] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:17:50.481 [2024-11-19 15:23:40.746708] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:17:50.481 [2024-11-19 15:23:40.746737] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:17:50.481 [2024-11-19 15:23:40.746833] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:17:50.481 [2024-11-19 15:23:40.746847] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:17:50.481 [2024-11-19 15:23:40.746863] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:17:50.481 [2024-11-19 15:23:40.746900] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:17:50.481 [2024-11-19 15:23:40.746955] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:17:50.481 [2024-11-19 15:23:40.746966] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:17:50.481 [2024-11-19 15:23:40.747060] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:17:50.481 [2024-11-19 15:23:40.747111] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:17:50.481 [2024-11-19 15:23:40.747118] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:17:50.481 [2024-11-19 15:23:40.747176] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:50.481 pt1 00:17:50.481 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:50.481 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:17:50.481 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:50.481 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:50.481 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:50.481 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:50.481 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:50.481 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:50.481 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:50.481 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:50.481 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:50.481 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:50.481 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:50.481 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:50.481 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:50.481 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:50.481 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:50.481 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:50.481 "name": "raid_bdev1", 00:17:50.481 "uuid": "f4aab9be-5639-4d45-a9be-7b917d7dc3dc", 00:17:50.481 "strip_size_kb": 0, 00:17:50.481 "state": "online", 00:17:50.481 "raid_level": "raid1", 00:17:50.481 "superblock": true, 00:17:50.481 "num_base_bdevs": 2, 00:17:50.481 "num_base_bdevs_discovered": 1, 00:17:50.481 "num_base_bdevs_operational": 1, 00:17:50.481 "base_bdevs_list": [ 00:17:50.481 { 00:17:50.481 "name": null, 00:17:50.481 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:50.481 "is_configured": false, 00:17:50.481 "data_offset": 256, 00:17:50.481 "data_size": 7936 00:17:50.481 }, 00:17:50.481 { 00:17:50.481 "name": "pt2", 00:17:50.481 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:50.481 "is_configured": true, 00:17:50.481 "data_offset": 256, 00:17:50.481 "data_size": 7936 00:17:50.481 } 00:17:50.481 ] 00:17:50.481 }' 00:17:50.481 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:50.481 15:23:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:51.052 15:23:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:17:51.052 15:23:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:51.052 15:23:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:51.052 15:23:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:17:51.052 15:23:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:51.052 15:23:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:17:51.052 15:23:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:17:51.052 15:23:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:51.052 15:23:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:51.052 15:23:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:17:51.052 [2024-11-19 15:23:41.264163] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:17:51.052 15:23:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:51.052 15:23:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@558 -- # '[' f4aab9be-5639-4d45-a9be-7b917d7dc3dc '!=' f4aab9be-5639-4d45-a9be-7b917d7dc3dc ']' 00:17:51.052 15:23:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@563 -- # killprocess 99031 00:17:51.052 15:23:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@954 -- # '[' -z 99031 ']' 00:17:51.052 15:23:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@958 -- # kill -0 99031 00:17:51.052 15:23:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@959 -- # uname 00:17:51.052 15:23:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:17:51.052 15:23:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 99031 00:17:51.052 15:23:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:17:51.052 15:23:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:17:51.052 15:23:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@972 -- # echo 'killing process with pid 99031' 00:17:51.052 killing process with pid 99031 00:17:51.052 15:23:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@973 -- # kill 99031 00:17:51.052 [2024-11-19 15:23:41.340910] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:17:51.052 [2024-11-19 15:23:41.341038] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:51.052 [2024-11-19 15:23:41.341105] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:51.052 15:23:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@978 -- # wait 99031 00:17:51.052 [2024-11-19 15:23:41.341149] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:17:51.052 [2024-11-19 15:23:41.364867] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:17:51.312 15:23:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@565 -- # return 0 00:17:51.312 00:17:51.312 real 0m4.885s 00:17:51.312 user 0m7.987s 00:17:51.312 sys 0m1.080s 00:17:51.312 15:23:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@1130 -- # xtrace_disable 00:17:51.312 15:23:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:51.312 ************************************ 00:17:51.312 END TEST raid_superblock_test_md_interleaved 00:17:51.312 ************************************ 00:17:51.312 15:23:41 bdev_raid -- bdev/bdev_raid.sh@1013 -- # run_test raid_rebuild_test_sb_md_interleaved raid_rebuild_test raid1 2 true false false 00:17:51.312 15:23:41 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:17:51.312 15:23:41 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:17:51.312 15:23:41 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:17:51.573 ************************************ 00:17:51.573 START TEST raid_rebuild_test_sb_md_interleaved 00:17:51.573 ************************************ 00:17:51.573 15:23:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 true false false 00:17:51.573 15:23:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:17:51.573 15:23:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:17:51.573 15:23:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:17:51.573 15:23:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:17:51.573 15:23:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@573 -- # local verify=false 00:17:51.573 15:23:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:17:51.573 15:23:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:17:51.573 15:23:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:17:51.573 15:23:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:17:51.573 15:23:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:17:51.573 15:23:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:17:51.573 15:23:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:17:51.573 15:23:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:17:51.573 15:23:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:17:51.573 15:23:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:17:51.573 15:23:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:17:51.573 15:23:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@576 -- # local strip_size 00:17:51.573 15:23:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@577 -- # local create_arg 00:17:51.573 15:23:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:17:51.573 15:23:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@579 -- # local data_offset 00:17:51.573 15:23:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:17:51.573 15:23:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:17:51.573 15:23:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:17:51.573 15:23:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:17:51.573 15:23:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@597 -- # raid_pid=99348 00:17:51.573 15:23:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:17:51.573 15:23:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@598 -- # waitforlisten 99348 00:17:51.573 15:23:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@835 -- # '[' -z 99348 ']' 00:17:51.573 15:23:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:17:51.573 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:17:51.573 15:23:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@840 -- # local max_retries=100 00:17:51.573 15:23:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:17:51.573 15:23:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@844 -- # xtrace_disable 00:17:51.573 15:23:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:51.573 I/O size of 3145728 is greater than zero copy threshold (65536). 00:17:51.573 Zero copy mechanism will not be used. 00:17:51.573 [2024-11-19 15:23:41.768788] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:17:51.573 [2024-11-19 15:23:41.768923] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid99348 ] 00:17:51.833 [2024-11-19 15:23:41.925021] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:51.833 [2024-11-19 15:23:41.949974] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:51.833 [2024-11-19 15:23:41.992736] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:17:51.833 [2024-11-19 15:23:41.992847] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:17:52.403 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:17:52.403 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@868 -- # return 0 00:17:52.403 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:17:52.403 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev1_malloc 00:17:52.403 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:52.403 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:52.403 BaseBdev1_malloc 00:17:52.403 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:52.403 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:17:52.403 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:52.403 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:52.403 [2024-11-19 15:23:42.594717] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:17:52.403 [2024-11-19 15:23:42.594787] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:52.403 [2024-11-19 15:23:42.594814] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:17:52.403 [2024-11-19 15:23:42.594822] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:52.403 [2024-11-19 15:23:42.596659] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:52.403 [2024-11-19 15:23:42.596696] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:17:52.403 BaseBdev1 00:17:52.403 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:52.403 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:17:52.403 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev2_malloc 00:17:52.403 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:52.404 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:52.404 BaseBdev2_malloc 00:17:52.404 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:52.404 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:17:52.404 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:52.404 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:52.404 [2024-11-19 15:23:42.623322] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:17:52.404 [2024-11-19 15:23:42.623442] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:52.404 [2024-11-19 15:23:42.623495] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:17:52.404 [2024-11-19 15:23:42.623524] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:52.404 [2024-11-19 15:23:42.625323] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:52.404 [2024-11-19 15:23:42.625413] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:17:52.404 BaseBdev2 00:17:52.404 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:52.404 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b spare_malloc 00:17:52.404 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:52.404 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:52.404 spare_malloc 00:17:52.404 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:52.404 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:17:52.404 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:52.404 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:52.404 spare_delay 00:17:52.404 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:52.404 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:17:52.404 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:52.404 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:52.404 [2024-11-19 15:23:42.663920] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:17:52.404 [2024-11-19 15:23:42.664042] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:52.404 [2024-11-19 15:23:42.664084] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:17:52.404 [2024-11-19 15:23:42.664112] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:52.404 [2024-11-19 15:23:42.665962] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:52.404 [2024-11-19 15:23:42.666035] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:17:52.404 spare 00:17:52.404 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:52.404 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:17:52.404 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:52.404 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:52.404 [2024-11-19 15:23:42.675938] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:17:52.404 [2024-11-19 15:23:42.677731] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:17:52.404 [2024-11-19 15:23:42.677936] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:17:52.404 [2024-11-19 15:23:42.677970] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:17:52.404 [2024-11-19 15:23:42.678088] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:17:52.404 [2024-11-19 15:23:42.678191] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:17:52.404 [2024-11-19 15:23:42.678234] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:17:52.404 [2024-11-19 15:23:42.678327] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:52.404 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:52.404 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:17:52.404 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:52.404 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:52.404 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:52.404 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:52.404 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:52.404 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:52.404 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:52.404 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:52.404 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:52.404 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:52.404 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:52.404 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:52.404 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:52.404 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:52.404 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:52.404 "name": "raid_bdev1", 00:17:52.404 "uuid": "e025bbfe-cc1c-4ee3-97ed-a695e0cd3e6d", 00:17:52.404 "strip_size_kb": 0, 00:17:52.404 "state": "online", 00:17:52.404 "raid_level": "raid1", 00:17:52.404 "superblock": true, 00:17:52.404 "num_base_bdevs": 2, 00:17:52.404 "num_base_bdevs_discovered": 2, 00:17:52.404 "num_base_bdevs_operational": 2, 00:17:52.404 "base_bdevs_list": [ 00:17:52.404 { 00:17:52.404 "name": "BaseBdev1", 00:17:52.404 "uuid": "17732bd0-447f-51b8-ba88-f5f516753486", 00:17:52.404 "is_configured": true, 00:17:52.404 "data_offset": 256, 00:17:52.404 "data_size": 7936 00:17:52.404 }, 00:17:52.404 { 00:17:52.404 "name": "BaseBdev2", 00:17:52.404 "uuid": "0bb42b7d-5683-528d-bf16-2cf8c9f3bdc3", 00:17:52.404 "is_configured": true, 00:17:52.404 "data_offset": 256, 00:17:52.404 "data_size": 7936 00:17:52.404 } 00:17:52.404 ] 00:17:52.404 }' 00:17:52.404 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:52.404 15:23:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:52.974 15:23:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:17:52.974 15:23:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:17:52.974 15:23:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:52.974 15:23:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:52.974 [2024-11-19 15:23:43.131353] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:17:52.974 15:23:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:52.974 15:23:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=7936 00:17:52.974 15:23:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:17:52.974 15:23:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:52.974 15:23:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:52.974 15:23:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:52.974 15:23:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:52.974 15:23:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@619 -- # data_offset=256 00:17:52.974 15:23:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:17:52.974 15:23:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@624 -- # '[' false = true ']' 00:17:52.974 15:23:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:17:52.974 15:23:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:52.975 15:23:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:52.975 [2024-11-19 15:23:43.214958] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:17:52.975 15:23:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:52.975 15:23:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:52.975 15:23:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:52.975 15:23:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:52.975 15:23:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:52.975 15:23:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:52.975 15:23:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:52.975 15:23:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:52.975 15:23:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:52.975 15:23:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:52.975 15:23:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:52.975 15:23:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:52.975 15:23:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:52.975 15:23:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:52.975 15:23:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:52.975 15:23:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:52.975 15:23:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:52.975 "name": "raid_bdev1", 00:17:52.975 "uuid": "e025bbfe-cc1c-4ee3-97ed-a695e0cd3e6d", 00:17:52.975 "strip_size_kb": 0, 00:17:52.975 "state": "online", 00:17:52.975 "raid_level": "raid1", 00:17:52.975 "superblock": true, 00:17:52.975 "num_base_bdevs": 2, 00:17:52.975 "num_base_bdevs_discovered": 1, 00:17:52.975 "num_base_bdevs_operational": 1, 00:17:52.975 "base_bdevs_list": [ 00:17:52.975 { 00:17:52.975 "name": null, 00:17:52.975 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:52.975 "is_configured": false, 00:17:52.975 "data_offset": 0, 00:17:52.975 "data_size": 7936 00:17:52.975 }, 00:17:52.975 { 00:17:52.975 "name": "BaseBdev2", 00:17:52.975 "uuid": "0bb42b7d-5683-528d-bf16-2cf8c9f3bdc3", 00:17:52.975 "is_configured": true, 00:17:52.975 "data_offset": 256, 00:17:52.975 "data_size": 7936 00:17:52.975 } 00:17:52.975 ] 00:17:52.975 }' 00:17:52.975 15:23:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:52.975 15:23:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:53.544 15:23:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:17:53.544 15:23:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:53.544 15:23:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:53.544 [2024-11-19 15:23:43.674165] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:17:53.544 [2024-11-19 15:23:43.693163] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:17:53.544 15:23:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:53.544 15:23:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@647 -- # sleep 1 00:17:53.544 [2024-11-19 15:23:43.700048] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:17:54.485 15:23:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:54.485 15:23:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:54.485 15:23:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:54.485 15:23:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:54.485 15:23:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:54.485 15:23:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:54.485 15:23:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:54.485 15:23:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:54.485 15:23:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:54.485 15:23:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:54.485 15:23:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:54.485 "name": "raid_bdev1", 00:17:54.485 "uuid": "e025bbfe-cc1c-4ee3-97ed-a695e0cd3e6d", 00:17:54.485 "strip_size_kb": 0, 00:17:54.485 "state": "online", 00:17:54.485 "raid_level": "raid1", 00:17:54.485 "superblock": true, 00:17:54.485 "num_base_bdevs": 2, 00:17:54.485 "num_base_bdevs_discovered": 2, 00:17:54.485 "num_base_bdevs_operational": 2, 00:17:54.485 "process": { 00:17:54.485 "type": "rebuild", 00:17:54.485 "target": "spare", 00:17:54.485 "progress": { 00:17:54.485 "blocks": 2560, 00:17:54.485 "percent": 32 00:17:54.485 } 00:17:54.485 }, 00:17:54.485 "base_bdevs_list": [ 00:17:54.485 { 00:17:54.485 "name": "spare", 00:17:54.485 "uuid": "145fc544-fc15-5310-a4dd-d404e9741b0a", 00:17:54.485 "is_configured": true, 00:17:54.485 "data_offset": 256, 00:17:54.485 "data_size": 7936 00:17:54.485 }, 00:17:54.485 { 00:17:54.485 "name": "BaseBdev2", 00:17:54.485 "uuid": "0bb42b7d-5683-528d-bf16-2cf8c9f3bdc3", 00:17:54.485 "is_configured": true, 00:17:54.485 "data_offset": 256, 00:17:54.485 "data_size": 7936 00:17:54.485 } 00:17:54.485 ] 00:17:54.485 }' 00:17:54.485 15:23:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:54.485 15:23:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:54.485 15:23:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:54.745 15:23:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:54.745 15:23:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:17:54.745 15:23:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:54.745 15:23:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:54.745 [2024-11-19 15:23:44.859867] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:54.745 [2024-11-19 15:23:44.905678] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:17:54.745 [2024-11-19 15:23:44.905780] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:54.745 [2024-11-19 15:23:44.905833] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:54.745 [2024-11-19 15:23:44.905864] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:17:54.745 15:23:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:54.745 15:23:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:54.745 15:23:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:54.745 15:23:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:54.745 15:23:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:54.745 15:23:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:54.745 15:23:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:54.745 15:23:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:54.745 15:23:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:54.745 15:23:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:54.745 15:23:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:54.745 15:23:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:54.745 15:23:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:54.745 15:23:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:54.745 15:23:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:54.745 15:23:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:54.745 15:23:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:54.745 "name": "raid_bdev1", 00:17:54.745 "uuid": "e025bbfe-cc1c-4ee3-97ed-a695e0cd3e6d", 00:17:54.745 "strip_size_kb": 0, 00:17:54.745 "state": "online", 00:17:54.745 "raid_level": "raid1", 00:17:54.745 "superblock": true, 00:17:54.745 "num_base_bdevs": 2, 00:17:54.745 "num_base_bdevs_discovered": 1, 00:17:54.745 "num_base_bdevs_operational": 1, 00:17:54.745 "base_bdevs_list": [ 00:17:54.745 { 00:17:54.745 "name": null, 00:17:54.745 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:54.745 "is_configured": false, 00:17:54.745 "data_offset": 0, 00:17:54.745 "data_size": 7936 00:17:54.745 }, 00:17:54.745 { 00:17:54.745 "name": "BaseBdev2", 00:17:54.745 "uuid": "0bb42b7d-5683-528d-bf16-2cf8c9f3bdc3", 00:17:54.745 "is_configured": true, 00:17:54.745 "data_offset": 256, 00:17:54.745 "data_size": 7936 00:17:54.745 } 00:17:54.745 ] 00:17:54.745 }' 00:17:54.745 15:23:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:54.745 15:23:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:55.316 15:23:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:55.316 15:23:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:55.316 15:23:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:55.316 15:23:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:55.316 15:23:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:55.316 15:23:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:55.316 15:23:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:55.316 15:23:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:55.316 15:23:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:55.316 15:23:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:55.316 15:23:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:55.316 "name": "raid_bdev1", 00:17:55.316 "uuid": "e025bbfe-cc1c-4ee3-97ed-a695e0cd3e6d", 00:17:55.316 "strip_size_kb": 0, 00:17:55.316 "state": "online", 00:17:55.316 "raid_level": "raid1", 00:17:55.316 "superblock": true, 00:17:55.316 "num_base_bdevs": 2, 00:17:55.316 "num_base_bdevs_discovered": 1, 00:17:55.316 "num_base_bdevs_operational": 1, 00:17:55.316 "base_bdevs_list": [ 00:17:55.316 { 00:17:55.316 "name": null, 00:17:55.316 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:55.316 "is_configured": false, 00:17:55.316 "data_offset": 0, 00:17:55.316 "data_size": 7936 00:17:55.316 }, 00:17:55.316 { 00:17:55.316 "name": "BaseBdev2", 00:17:55.316 "uuid": "0bb42b7d-5683-528d-bf16-2cf8c9f3bdc3", 00:17:55.316 "is_configured": true, 00:17:55.316 "data_offset": 256, 00:17:55.316 "data_size": 7936 00:17:55.316 } 00:17:55.316 ] 00:17:55.316 }' 00:17:55.316 15:23:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:55.316 15:23:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:55.316 15:23:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:55.316 15:23:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:55.316 15:23:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:17:55.316 15:23:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:55.316 15:23:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:55.316 [2024-11-19 15:23:45.492670] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:17:55.316 [2024-11-19 15:23:45.496262] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:17:55.316 [2024-11-19 15:23:45.498096] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:17:55.316 15:23:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:55.316 15:23:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@663 -- # sleep 1 00:17:56.255 15:23:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:56.255 15:23:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:56.255 15:23:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:56.255 15:23:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:56.255 15:23:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:56.255 15:23:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:56.255 15:23:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:56.255 15:23:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:56.255 15:23:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:56.255 15:23:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:56.255 15:23:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:56.255 "name": "raid_bdev1", 00:17:56.255 "uuid": "e025bbfe-cc1c-4ee3-97ed-a695e0cd3e6d", 00:17:56.255 "strip_size_kb": 0, 00:17:56.255 "state": "online", 00:17:56.255 "raid_level": "raid1", 00:17:56.255 "superblock": true, 00:17:56.255 "num_base_bdevs": 2, 00:17:56.255 "num_base_bdevs_discovered": 2, 00:17:56.255 "num_base_bdevs_operational": 2, 00:17:56.255 "process": { 00:17:56.255 "type": "rebuild", 00:17:56.255 "target": "spare", 00:17:56.255 "progress": { 00:17:56.255 "blocks": 2560, 00:17:56.255 "percent": 32 00:17:56.255 } 00:17:56.255 }, 00:17:56.255 "base_bdevs_list": [ 00:17:56.255 { 00:17:56.255 "name": "spare", 00:17:56.255 "uuid": "145fc544-fc15-5310-a4dd-d404e9741b0a", 00:17:56.255 "is_configured": true, 00:17:56.255 "data_offset": 256, 00:17:56.255 "data_size": 7936 00:17:56.255 }, 00:17:56.255 { 00:17:56.255 "name": "BaseBdev2", 00:17:56.255 "uuid": "0bb42b7d-5683-528d-bf16-2cf8c9f3bdc3", 00:17:56.255 "is_configured": true, 00:17:56.255 "data_offset": 256, 00:17:56.255 "data_size": 7936 00:17:56.255 } 00:17:56.255 ] 00:17:56.255 }' 00:17:56.255 15:23:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:56.516 15:23:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:56.516 15:23:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:56.516 15:23:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:56.516 15:23:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:17:56.516 15:23:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:17:56.516 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:17:56.516 15:23:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:17:56.516 15:23:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:17:56.516 15:23:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:17:56.516 15:23:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@706 -- # local timeout=624 00:17:56.516 15:23:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:17:56.516 15:23:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:56.516 15:23:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:56.516 15:23:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:56.516 15:23:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:56.516 15:23:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:56.516 15:23:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:56.516 15:23:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:56.516 15:23:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:56.516 15:23:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:56.516 15:23:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:56.516 15:23:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:56.516 "name": "raid_bdev1", 00:17:56.516 "uuid": "e025bbfe-cc1c-4ee3-97ed-a695e0cd3e6d", 00:17:56.516 "strip_size_kb": 0, 00:17:56.516 "state": "online", 00:17:56.516 "raid_level": "raid1", 00:17:56.516 "superblock": true, 00:17:56.516 "num_base_bdevs": 2, 00:17:56.516 "num_base_bdevs_discovered": 2, 00:17:56.516 "num_base_bdevs_operational": 2, 00:17:56.516 "process": { 00:17:56.516 "type": "rebuild", 00:17:56.516 "target": "spare", 00:17:56.516 "progress": { 00:17:56.516 "blocks": 2816, 00:17:56.516 "percent": 35 00:17:56.516 } 00:17:56.516 }, 00:17:56.516 "base_bdevs_list": [ 00:17:56.516 { 00:17:56.516 "name": "spare", 00:17:56.516 "uuid": "145fc544-fc15-5310-a4dd-d404e9741b0a", 00:17:56.516 "is_configured": true, 00:17:56.516 "data_offset": 256, 00:17:56.516 "data_size": 7936 00:17:56.516 }, 00:17:56.516 { 00:17:56.516 "name": "BaseBdev2", 00:17:56.516 "uuid": "0bb42b7d-5683-528d-bf16-2cf8c9f3bdc3", 00:17:56.516 "is_configured": true, 00:17:56.516 "data_offset": 256, 00:17:56.516 "data_size": 7936 00:17:56.516 } 00:17:56.516 ] 00:17:56.516 }' 00:17:56.516 15:23:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:56.516 15:23:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:56.516 15:23:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:56.516 15:23:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:56.516 15:23:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@711 -- # sleep 1 00:17:57.899 15:23:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:17:57.899 15:23:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:57.899 15:23:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:57.899 15:23:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:57.899 15:23:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:57.899 15:23:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:57.899 15:23:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:57.899 15:23:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:57.899 15:23:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:57.899 15:23:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:57.899 15:23:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:57.899 15:23:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:57.899 "name": "raid_bdev1", 00:17:57.899 "uuid": "e025bbfe-cc1c-4ee3-97ed-a695e0cd3e6d", 00:17:57.899 "strip_size_kb": 0, 00:17:57.899 "state": "online", 00:17:57.899 "raid_level": "raid1", 00:17:57.899 "superblock": true, 00:17:57.899 "num_base_bdevs": 2, 00:17:57.899 "num_base_bdevs_discovered": 2, 00:17:57.899 "num_base_bdevs_operational": 2, 00:17:57.899 "process": { 00:17:57.899 "type": "rebuild", 00:17:57.899 "target": "spare", 00:17:57.899 "progress": { 00:17:57.899 "blocks": 5888, 00:17:57.899 "percent": 74 00:17:57.899 } 00:17:57.899 }, 00:17:57.899 "base_bdevs_list": [ 00:17:57.899 { 00:17:57.899 "name": "spare", 00:17:57.899 "uuid": "145fc544-fc15-5310-a4dd-d404e9741b0a", 00:17:57.899 "is_configured": true, 00:17:57.899 "data_offset": 256, 00:17:57.899 "data_size": 7936 00:17:57.899 }, 00:17:57.899 { 00:17:57.899 "name": "BaseBdev2", 00:17:57.899 "uuid": "0bb42b7d-5683-528d-bf16-2cf8c9f3bdc3", 00:17:57.899 "is_configured": true, 00:17:57.899 "data_offset": 256, 00:17:57.899 "data_size": 7936 00:17:57.899 } 00:17:57.899 ] 00:17:57.899 }' 00:17:57.899 15:23:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:57.899 15:23:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:57.899 15:23:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:57.899 15:23:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:57.899 15:23:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@711 -- # sleep 1 00:17:58.468 [2024-11-19 15:23:48.608448] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:17:58.468 [2024-11-19 15:23:48.608521] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:17:58.468 [2024-11-19 15:23:48.608648] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:58.728 15:23:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:17:58.728 15:23:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:58.728 15:23:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:58.728 15:23:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:58.728 15:23:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:58.728 15:23:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:58.728 15:23:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:58.728 15:23:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:58.728 15:23:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:58.728 15:23:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:58.728 15:23:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:58.728 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:58.728 "name": "raid_bdev1", 00:17:58.728 "uuid": "e025bbfe-cc1c-4ee3-97ed-a695e0cd3e6d", 00:17:58.728 "strip_size_kb": 0, 00:17:58.728 "state": "online", 00:17:58.728 "raid_level": "raid1", 00:17:58.728 "superblock": true, 00:17:58.728 "num_base_bdevs": 2, 00:17:58.728 "num_base_bdevs_discovered": 2, 00:17:58.728 "num_base_bdevs_operational": 2, 00:17:58.728 "base_bdevs_list": [ 00:17:58.728 { 00:17:58.728 "name": "spare", 00:17:58.728 "uuid": "145fc544-fc15-5310-a4dd-d404e9741b0a", 00:17:58.728 "is_configured": true, 00:17:58.728 "data_offset": 256, 00:17:58.728 "data_size": 7936 00:17:58.728 }, 00:17:58.728 { 00:17:58.728 "name": "BaseBdev2", 00:17:58.728 "uuid": "0bb42b7d-5683-528d-bf16-2cf8c9f3bdc3", 00:17:58.728 "is_configured": true, 00:17:58.728 "data_offset": 256, 00:17:58.728 "data_size": 7936 00:17:58.728 } 00:17:58.728 ] 00:17:58.728 }' 00:17:58.728 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:58.728 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:17:58.728 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:58.988 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:17:58.988 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@709 -- # break 00:17:58.988 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:58.988 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:58.988 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:58.988 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:58.988 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:58.988 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:58.988 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:58.988 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:58.988 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:58.988 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:58.988 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:58.988 "name": "raid_bdev1", 00:17:58.988 "uuid": "e025bbfe-cc1c-4ee3-97ed-a695e0cd3e6d", 00:17:58.988 "strip_size_kb": 0, 00:17:58.988 "state": "online", 00:17:58.988 "raid_level": "raid1", 00:17:58.988 "superblock": true, 00:17:58.988 "num_base_bdevs": 2, 00:17:58.988 "num_base_bdevs_discovered": 2, 00:17:58.988 "num_base_bdevs_operational": 2, 00:17:58.988 "base_bdevs_list": [ 00:17:58.988 { 00:17:58.988 "name": "spare", 00:17:58.988 "uuid": "145fc544-fc15-5310-a4dd-d404e9741b0a", 00:17:58.988 "is_configured": true, 00:17:58.988 "data_offset": 256, 00:17:58.988 "data_size": 7936 00:17:58.988 }, 00:17:58.988 { 00:17:58.988 "name": "BaseBdev2", 00:17:58.988 "uuid": "0bb42b7d-5683-528d-bf16-2cf8c9f3bdc3", 00:17:58.988 "is_configured": true, 00:17:58.988 "data_offset": 256, 00:17:58.988 "data_size": 7936 00:17:58.988 } 00:17:58.988 ] 00:17:58.988 }' 00:17:58.988 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:58.988 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:58.988 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:58.988 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:58.988 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:17:58.988 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:58.988 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:58.988 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:58.988 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:58.988 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:58.988 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:58.988 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:58.988 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:58.988 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:58.988 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:58.988 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:58.988 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:58.988 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:58.988 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:58.988 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:58.988 "name": "raid_bdev1", 00:17:58.988 "uuid": "e025bbfe-cc1c-4ee3-97ed-a695e0cd3e6d", 00:17:58.988 "strip_size_kb": 0, 00:17:58.988 "state": "online", 00:17:58.988 "raid_level": "raid1", 00:17:58.988 "superblock": true, 00:17:58.988 "num_base_bdevs": 2, 00:17:58.988 "num_base_bdevs_discovered": 2, 00:17:58.988 "num_base_bdevs_operational": 2, 00:17:58.988 "base_bdevs_list": [ 00:17:58.988 { 00:17:58.988 "name": "spare", 00:17:58.988 "uuid": "145fc544-fc15-5310-a4dd-d404e9741b0a", 00:17:58.988 "is_configured": true, 00:17:58.988 "data_offset": 256, 00:17:58.988 "data_size": 7936 00:17:58.988 }, 00:17:58.988 { 00:17:58.989 "name": "BaseBdev2", 00:17:58.989 "uuid": "0bb42b7d-5683-528d-bf16-2cf8c9f3bdc3", 00:17:58.989 "is_configured": true, 00:17:58.989 "data_offset": 256, 00:17:58.989 "data_size": 7936 00:17:58.989 } 00:17:58.989 ] 00:17:58.989 }' 00:17:58.989 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:58.989 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:59.559 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:17:59.559 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:59.559 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:59.559 [2024-11-19 15:23:49.666636] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:17:59.559 [2024-11-19 15:23:49.666725] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:17:59.559 [2024-11-19 15:23:49.666841] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:59.559 [2024-11-19 15:23:49.666927] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:59.559 [2024-11-19 15:23:49.666992] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:17:59.559 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:59.559 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:59.559 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@720 -- # jq length 00:17:59.559 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:59.559 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:59.559 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:59.559 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:17:59.559 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@722 -- # '[' false = true ']' 00:17:59.559 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:17:59.559 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:17:59.559 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:59.559 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:59.559 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:59.559 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:17:59.559 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:59.559 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:59.559 [2024-11-19 15:23:49.738509] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:17:59.559 [2024-11-19 15:23:49.738576] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:59.559 [2024-11-19 15:23:49.738598] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:17:59.559 [2024-11-19 15:23:49.738609] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:59.559 [2024-11-19 15:23:49.740533] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:59.559 [2024-11-19 15:23:49.740573] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:17:59.559 [2024-11-19 15:23:49.740623] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:17:59.559 [2024-11-19 15:23:49.740668] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:17:59.559 [2024-11-19 15:23:49.740750] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:17:59.559 spare 00:17:59.559 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:59.559 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:17:59.559 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:59.559 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:59.559 [2024-11-19 15:23:49.840639] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:17:59.559 [2024-11-19 15:23:49.840663] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:17:59.559 [2024-11-19 15:23:49.840754] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:17:59.559 [2024-11-19 15:23:49.840831] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:17:59.559 [2024-11-19 15:23:49.840843] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:17:59.559 [2024-11-19 15:23:49.840913] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:59.559 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:59.559 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:17:59.559 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:59.559 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:59.559 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:59.559 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:59.559 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:59.559 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:59.559 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:59.559 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:59.559 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:59.559 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:59.559 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:59.559 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:59.559 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:59.559 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:59.559 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:59.559 "name": "raid_bdev1", 00:17:59.559 "uuid": "e025bbfe-cc1c-4ee3-97ed-a695e0cd3e6d", 00:17:59.559 "strip_size_kb": 0, 00:17:59.559 "state": "online", 00:17:59.559 "raid_level": "raid1", 00:17:59.559 "superblock": true, 00:17:59.559 "num_base_bdevs": 2, 00:17:59.559 "num_base_bdevs_discovered": 2, 00:17:59.559 "num_base_bdevs_operational": 2, 00:17:59.559 "base_bdevs_list": [ 00:17:59.559 { 00:17:59.559 "name": "spare", 00:17:59.559 "uuid": "145fc544-fc15-5310-a4dd-d404e9741b0a", 00:17:59.559 "is_configured": true, 00:17:59.559 "data_offset": 256, 00:17:59.559 "data_size": 7936 00:17:59.559 }, 00:17:59.559 { 00:17:59.559 "name": "BaseBdev2", 00:17:59.559 "uuid": "0bb42b7d-5683-528d-bf16-2cf8c9f3bdc3", 00:17:59.559 "is_configured": true, 00:17:59.559 "data_offset": 256, 00:17:59.559 "data_size": 7936 00:17:59.559 } 00:17:59.559 ] 00:17:59.559 }' 00:17:59.559 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:59.559 15:23:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:18:00.130 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:18:00.130 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:18:00.130 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:18:00.130 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:18:00.130 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:18:00.130 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:18:00.130 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:00.130 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:18:00.130 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:18:00.130 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:00.130 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:18:00.130 "name": "raid_bdev1", 00:18:00.130 "uuid": "e025bbfe-cc1c-4ee3-97ed-a695e0cd3e6d", 00:18:00.130 "strip_size_kb": 0, 00:18:00.130 "state": "online", 00:18:00.130 "raid_level": "raid1", 00:18:00.130 "superblock": true, 00:18:00.130 "num_base_bdevs": 2, 00:18:00.130 "num_base_bdevs_discovered": 2, 00:18:00.130 "num_base_bdevs_operational": 2, 00:18:00.130 "base_bdevs_list": [ 00:18:00.130 { 00:18:00.130 "name": "spare", 00:18:00.130 "uuid": "145fc544-fc15-5310-a4dd-d404e9741b0a", 00:18:00.130 "is_configured": true, 00:18:00.130 "data_offset": 256, 00:18:00.130 "data_size": 7936 00:18:00.130 }, 00:18:00.130 { 00:18:00.130 "name": "BaseBdev2", 00:18:00.130 "uuid": "0bb42b7d-5683-528d-bf16-2cf8c9f3bdc3", 00:18:00.130 "is_configured": true, 00:18:00.130 "data_offset": 256, 00:18:00.130 "data_size": 7936 00:18:00.130 } 00:18:00.130 ] 00:18:00.130 }' 00:18:00.130 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:18:00.130 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:18:00.130 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:18:00.390 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:18:00.390 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:18:00.390 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:00.390 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:18:00.390 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:18:00.390 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:00.390 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:18:00.390 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:18:00.390 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:00.390 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:18:00.390 [2024-11-19 15:23:50.517234] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:18:00.390 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:00.390 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:18:00.390 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:18:00.390 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:18:00.390 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:18:00.390 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:18:00.390 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:18:00.390 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:18:00.390 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:18:00.390 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:18:00.390 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:18:00.390 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:18:00.390 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:18:00.390 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:00.390 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:18:00.390 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:00.390 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:18:00.390 "name": "raid_bdev1", 00:18:00.390 "uuid": "e025bbfe-cc1c-4ee3-97ed-a695e0cd3e6d", 00:18:00.390 "strip_size_kb": 0, 00:18:00.390 "state": "online", 00:18:00.390 "raid_level": "raid1", 00:18:00.390 "superblock": true, 00:18:00.390 "num_base_bdevs": 2, 00:18:00.390 "num_base_bdevs_discovered": 1, 00:18:00.390 "num_base_bdevs_operational": 1, 00:18:00.390 "base_bdevs_list": [ 00:18:00.390 { 00:18:00.390 "name": null, 00:18:00.390 "uuid": "00000000-0000-0000-0000-000000000000", 00:18:00.390 "is_configured": false, 00:18:00.390 "data_offset": 0, 00:18:00.390 "data_size": 7936 00:18:00.390 }, 00:18:00.390 { 00:18:00.390 "name": "BaseBdev2", 00:18:00.390 "uuid": "0bb42b7d-5683-528d-bf16-2cf8c9f3bdc3", 00:18:00.390 "is_configured": true, 00:18:00.390 "data_offset": 256, 00:18:00.390 "data_size": 7936 00:18:00.390 } 00:18:00.390 ] 00:18:00.390 }' 00:18:00.390 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:18:00.390 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:18:00.650 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:18:00.650 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:00.650 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:18:00.650 [2024-11-19 15:23:50.924568] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:18:00.650 [2024-11-19 15:23:50.924787] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:18:00.650 [2024-11-19 15:23:50.924873] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:18:00.650 [2024-11-19 15:23:50.924941] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:18:00.650 [2024-11-19 15:23:50.928425] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:18:00.651 [2024-11-19 15:23:50.930301] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:18:00.651 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:00.651 15:23:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@757 -- # sleep 1 00:18:02.032 15:23:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:18:02.032 15:23:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:18:02.032 15:23:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:18:02.032 15:23:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:18:02.032 15:23:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:18:02.032 15:23:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:18:02.032 15:23:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:18:02.032 15:23:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:02.032 15:23:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:18:02.032 15:23:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:02.032 15:23:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:18:02.032 "name": "raid_bdev1", 00:18:02.032 "uuid": "e025bbfe-cc1c-4ee3-97ed-a695e0cd3e6d", 00:18:02.032 "strip_size_kb": 0, 00:18:02.032 "state": "online", 00:18:02.032 "raid_level": "raid1", 00:18:02.032 "superblock": true, 00:18:02.032 "num_base_bdevs": 2, 00:18:02.032 "num_base_bdevs_discovered": 2, 00:18:02.032 "num_base_bdevs_operational": 2, 00:18:02.032 "process": { 00:18:02.032 "type": "rebuild", 00:18:02.032 "target": "spare", 00:18:02.032 "progress": { 00:18:02.032 "blocks": 2560, 00:18:02.032 "percent": 32 00:18:02.032 } 00:18:02.032 }, 00:18:02.032 "base_bdevs_list": [ 00:18:02.032 { 00:18:02.032 "name": "spare", 00:18:02.032 "uuid": "145fc544-fc15-5310-a4dd-d404e9741b0a", 00:18:02.032 "is_configured": true, 00:18:02.032 "data_offset": 256, 00:18:02.032 "data_size": 7936 00:18:02.032 }, 00:18:02.032 { 00:18:02.032 "name": "BaseBdev2", 00:18:02.032 "uuid": "0bb42b7d-5683-528d-bf16-2cf8c9f3bdc3", 00:18:02.032 "is_configured": true, 00:18:02.032 "data_offset": 256, 00:18:02.032 "data_size": 7936 00:18:02.032 } 00:18:02.032 ] 00:18:02.032 }' 00:18:02.032 15:23:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:18:02.032 15:23:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:18:02.032 15:23:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:18:02.032 15:23:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:18:02.032 15:23:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:18:02.032 15:23:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:02.032 15:23:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:18:02.032 [2024-11-19 15:23:52.096979] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:18:02.032 [2024-11-19 15:23:52.134148] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:18:02.032 [2024-11-19 15:23:52.134197] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:18:02.032 [2024-11-19 15:23:52.134213] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:18:02.032 [2024-11-19 15:23:52.134219] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:18:02.032 15:23:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:02.032 15:23:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:18:02.032 15:23:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:18:02.032 15:23:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:18:02.032 15:23:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:18:02.032 15:23:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:18:02.032 15:23:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:18:02.032 15:23:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:18:02.032 15:23:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:18:02.032 15:23:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:18:02.032 15:23:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:18:02.032 15:23:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:18:02.032 15:23:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:02.032 15:23:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:18:02.032 15:23:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:18:02.032 15:23:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:02.032 15:23:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:18:02.032 "name": "raid_bdev1", 00:18:02.032 "uuid": "e025bbfe-cc1c-4ee3-97ed-a695e0cd3e6d", 00:18:02.032 "strip_size_kb": 0, 00:18:02.032 "state": "online", 00:18:02.032 "raid_level": "raid1", 00:18:02.032 "superblock": true, 00:18:02.032 "num_base_bdevs": 2, 00:18:02.032 "num_base_bdevs_discovered": 1, 00:18:02.032 "num_base_bdevs_operational": 1, 00:18:02.032 "base_bdevs_list": [ 00:18:02.032 { 00:18:02.032 "name": null, 00:18:02.032 "uuid": "00000000-0000-0000-0000-000000000000", 00:18:02.032 "is_configured": false, 00:18:02.032 "data_offset": 0, 00:18:02.032 "data_size": 7936 00:18:02.032 }, 00:18:02.032 { 00:18:02.032 "name": "BaseBdev2", 00:18:02.032 "uuid": "0bb42b7d-5683-528d-bf16-2cf8c9f3bdc3", 00:18:02.032 "is_configured": true, 00:18:02.033 "data_offset": 256, 00:18:02.033 "data_size": 7936 00:18:02.033 } 00:18:02.033 ] 00:18:02.033 }' 00:18:02.033 15:23:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:18:02.033 15:23:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:18:02.292 15:23:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:18:02.292 15:23:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:02.292 15:23:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:18:02.292 [2024-11-19 15:23:52.593241] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:18:02.292 [2024-11-19 15:23:52.593341] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:18:02.292 [2024-11-19 15:23:52.593386] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:18:02.292 [2024-11-19 15:23:52.593405] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:18:02.292 [2024-11-19 15:23:52.593573] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:18:02.292 [2024-11-19 15:23:52.593584] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:18:02.292 [2024-11-19 15:23:52.593634] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:18:02.292 [2024-11-19 15:23:52.593645] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:18:02.292 [2024-11-19 15:23:52.593665] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:18:02.292 [2024-11-19 15:23:52.593695] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:18:02.292 [2024-11-19 15:23:52.596256] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:18:02.292 [2024-11-19 15:23:52.598143] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:18:02.292 spare 00:18:02.292 15:23:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:02.292 15:23:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@764 -- # sleep 1 00:18:03.682 15:23:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:18:03.682 15:23:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:18:03.682 15:23:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:18:03.682 15:23:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:18:03.682 15:23:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:18:03.682 15:23:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:18:03.682 15:23:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:18:03.682 15:23:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:03.683 15:23:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:18:03.683 15:23:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:03.683 15:23:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:18:03.683 "name": "raid_bdev1", 00:18:03.683 "uuid": "e025bbfe-cc1c-4ee3-97ed-a695e0cd3e6d", 00:18:03.683 "strip_size_kb": 0, 00:18:03.683 "state": "online", 00:18:03.683 "raid_level": "raid1", 00:18:03.683 "superblock": true, 00:18:03.683 "num_base_bdevs": 2, 00:18:03.683 "num_base_bdevs_discovered": 2, 00:18:03.683 "num_base_bdevs_operational": 2, 00:18:03.683 "process": { 00:18:03.683 "type": "rebuild", 00:18:03.683 "target": "spare", 00:18:03.683 "progress": { 00:18:03.683 "blocks": 2560, 00:18:03.683 "percent": 32 00:18:03.683 } 00:18:03.683 }, 00:18:03.683 "base_bdevs_list": [ 00:18:03.683 { 00:18:03.683 "name": "spare", 00:18:03.683 "uuid": "145fc544-fc15-5310-a4dd-d404e9741b0a", 00:18:03.683 "is_configured": true, 00:18:03.683 "data_offset": 256, 00:18:03.683 "data_size": 7936 00:18:03.683 }, 00:18:03.683 { 00:18:03.683 "name": "BaseBdev2", 00:18:03.683 "uuid": "0bb42b7d-5683-528d-bf16-2cf8c9f3bdc3", 00:18:03.683 "is_configured": true, 00:18:03.683 "data_offset": 256, 00:18:03.683 "data_size": 7936 00:18:03.683 } 00:18:03.683 ] 00:18:03.683 }' 00:18:03.683 15:23:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:18:03.683 15:23:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:18:03.683 15:23:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:18:03.683 15:23:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:18:03.683 15:23:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:18:03.683 15:23:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:03.683 15:23:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:18:03.683 [2024-11-19 15:23:53.760778] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:18:03.683 [2024-11-19 15:23:53.801955] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:18:03.683 [2024-11-19 15:23:53.802040] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:18:03.683 [2024-11-19 15:23:53.802053] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:18:03.683 [2024-11-19 15:23:53.802062] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:18:03.683 15:23:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:03.683 15:23:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:18:03.683 15:23:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:18:03.683 15:23:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:18:03.683 15:23:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:18:03.683 15:23:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:18:03.683 15:23:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:18:03.683 15:23:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:18:03.683 15:23:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:18:03.683 15:23:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:18:03.683 15:23:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:18:03.683 15:23:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:18:03.683 15:23:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:03.683 15:23:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:18:03.683 15:23:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:18:03.683 15:23:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:03.683 15:23:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:18:03.683 "name": "raid_bdev1", 00:18:03.683 "uuid": "e025bbfe-cc1c-4ee3-97ed-a695e0cd3e6d", 00:18:03.683 "strip_size_kb": 0, 00:18:03.683 "state": "online", 00:18:03.683 "raid_level": "raid1", 00:18:03.683 "superblock": true, 00:18:03.683 "num_base_bdevs": 2, 00:18:03.683 "num_base_bdevs_discovered": 1, 00:18:03.683 "num_base_bdevs_operational": 1, 00:18:03.683 "base_bdevs_list": [ 00:18:03.683 { 00:18:03.683 "name": null, 00:18:03.683 "uuid": "00000000-0000-0000-0000-000000000000", 00:18:03.683 "is_configured": false, 00:18:03.683 "data_offset": 0, 00:18:03.683 "data_size": 7936 00:18:03.683 }, 00:18:03.683 { 00:18:03.683 "name": "BaseBdev2", 00:18:03.683 "uuid": "0bb42b7d-5683-528d-bf16-2cf8c9f3bdc3", 00:18:03.683 "is_configured": true, 00:18:03.683 "data_offset": 256, 00:18:03.683 "data_size": 7936 00:18:03.683 } 00:18:03.683 ] 00:18:03.683 }' 00:18:03.683 15:23:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:18:03.683 15:23:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:18:03.943 15:23:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:18:03.943 15:23:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:18:03.943 15:23:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:18:03.943 15:23:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:18:03.943 15:23:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:18:03.943 15:23:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:18:03.943 15:23:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:03.943 15:23:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:18:03.943 15:23:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:18:03.943 15:23:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:04.204 15:23:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:18:04.204 "name": "raid_bdev1", 00:18:04.204 "uuid": "e025bbfe-cc1c-4ee3-97ed-a695e0cd3e6d", 00:18:04.204 "strip_size_kb": 0, 00:18:04.204 "state": "online", 00:18:04.204 "raid_level": "raid1", 00:18:04.204 "superblock": true, 00:18:04.204 "num_base_bdevs": 2, 00:18:04.204 "num_base_bdevs_discovered": 1, 00:18:04.204 "num_base_bdevs_operational": 1, 00:18:04.204 "base_bdevs_list": [ 00:18:04.204 { 00:18:04.204 "name": null, 00:18:04.204 "uuid": "00000000-0000-0000-0000-000000000000", 00:18:04.204 "is_configured": false, 00:18:04.204 "data_offset": 0, 00:18:04.204 "data_size": 7936 00:18:04.204 }, 00:18:04.204 { 00:18:04.204 "name": "BaseBdev2", 00:18:04.204 "uuid": "0bb42b7d-5683-528d-bf16-2cf8c9f3bdc3", 00:18:04.204 "is_configured": true, 00:18:04.204 "data_offset": 256, 00:18:04.204 "data_size": 7936 00:18:04.204 } 00:18:04.204 ] 00:18:04.204 }' 00:18:04.204 15:23:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:18:04.204 15:23:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:18:04.204 15:23:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:18:04.204 15:23:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:18:04.204 15:23:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:18:04.204 15:23:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:04.204 15:23:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:18:04.204 15:23:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:04.204 15:23:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:18:04.204 15:23:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:04.204 15:23:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:18:04.204 [2024-11-19 15:23:54.356851] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:18:04.204 [2024-11-19 15:23:54.356905] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:18:04.204 [2024-11-19 15:23:54.356939] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:18:04.204 [2024-11-19 15:23:54.356949] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:18:04.204 [2024-11-19 15:23:54.357114] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:18:04.204 [2024-11-19 15:23:54.357130] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:18:04.204 [2024-11-19 15:23:54.357172] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:18:04.204 [2024-11-19 15:23:54.357187] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:18:04.204 [2024-11-19 15:23:54.357195] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:18:04.204 [2024-11-19 15:23:54.357207] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:18:04.204 BaseBdev1 00:18:04.204 15:23:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:04.204 15:23:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@775 -- # sleep 1 00:18:05.142 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:18:05.142 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:18:05.142 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:18:05.142 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:18:05.142 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:18:05.142 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:18:05.142 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:18:05.142 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:18:05.142 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:18:05.142 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:18:05.142 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:18:05.142 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:18:05.142 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:05.142 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:18:05.142 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:05.142 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:18:05.142 "name": "raid_bdev1", 00:18:05.142 "uuid": "e025bbfe-cc1c-4ee3-97ed-a695e0cd3e6d", 00:18:05.142 "strip_size_kb": 0, 00:18:05.143 "state": "online", 00:18:05.143 "raid_level": "raid1", 00:18:05.143 "superblock": true, 00:18:05.143 "num_base_bdevs": 2, 00:18:05.143 "num_base_bdevs_discovered": 1, 00:18:05.143 "num_base_bdevs_operational": 1, 00:18:05.143 "base_bdevs_list": [ 00:18:05.143 { 00:18:05.143 "name": null, 00:18:05.143 "uuid": "00000000-0000-0000-0000-000000000000", 00:18:05.143 "is_configured": false, 00:18:05.143 "data_offset": 0, 00:18:05.143 "data_size": 7936 00:18:05.143 }, 00:18:05.143 { 00:18:05.143 "name": "BaseBdev2", 00:18:05.143 "uuid": "0bb42b7d-5683-528d-bf16-2cf8c9f3bdc3", 00:18:05.143 "is_configured": true, 00:18:05.143 "data_offset": 256, 00:18:05.143 "data_size": 7936 00:18:05.143 } 00:18:05.143 ] 00:18:05.143 }' 00:18:05.143 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:18:05.143 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:18:05.712 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:18:05.712 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:18:05.712 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:18:05.712 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:18:05.712 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:18:05.712 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:18:05.712 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:18:05.712 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:05.712 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:18:05.712 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:05.712 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:18:05.712 "name": "raid_bdev1", 00:18:05.712 "uuid": "e025bbfe-cc1c-4ee3-97ed-a695e0cd3e6d", 00:18:05.712 "strip_size_kb": 0, 00:18:05.712 "state": "online", 00:18:05.712 "raid_level": "raid1", 00:18:05.712 "superblock": true, 00:18:05.712 "num_base_bdevs": 2, 00:18:05.712 "num_base_bdevs_discovered": 1, 00:18:05.712 "num_base_bdevs_operational": 1, 00:18:05.712 "base_bdevs_list": [ 00:18:05.712 { 00:18:05.712 "name": null, 00:18:05.712 "uuid": "00000000-0000-0000-0000-000000000000", 00:18:05.712 "is_configured": false, 00:18:05.712 "data_offset": 0, 00:18:05.712 "data_size": 7936 00:18:05.712 }, 00:18:05.712 { 00:18:05.712 "name": "BaseBdev2", 00:18:05.712 "uuid": "0bb42b7d-5683-528d-bf16-2cf8c9f3bdc3", 00:18:05.712 "is_configured": true, 00:18:05.712 "data_offset": 256, 00:18:05.712 "data_size": 7936 00:18:05.712 } 00:18:05.712 ] 00:18:05.712 }' 00:18:05.712 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:18:05.712 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:18:05.712 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:18:05.712 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:18:05.712 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:18:05.712 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@652 -- # local es=0 00:18:05.712 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:18:05.712 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:18:05.712 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:18:05.712 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:18:05.712 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:18:05.712 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:18:05.712 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:05.713 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:18:05.713 [2024-11-19 15:23:55.926236] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:18:05.713 [2024-11-19 15:23:55.926435] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:18:05.713 [2024-11-19 15:23:55.926467] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:18:05.713 request: 00:18:05.713 { 00:18:05.713 "base_bdev": "BaseBdev1", 00:18:05.713 "raid_bdev": "raid_bdev1", 00:18:05.713 "method": "bdev_raid_add_base_bdev", 00:18:05.713 "req_id": 1 00:18:05.713 } 00:18:05.713 Got JSON-RPC error response 00:18:05.713 response: 00:18:05.713 { 00:18:05.713 "code": -22, 00:18:05.713 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:18:05.713 } 00:18:05.713 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:18:05.713 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@655 -- # es=1 00:18:05.713 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:18:05.713 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:18:05.713 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:18:05.713 15:23:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@779 -- # sleep 1 00:18:06.653 15:23:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:18:06.653 15:23:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:18:06.653 15:23:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:18:06.653 15:23:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:18:06.653 15:23:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:18:06.653 15:23:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:18:06.653 15:23:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:18:06.653 15:23:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:18:06.653 15:23:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:18:06.653 15:23:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:18:06.653 15:23:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:18:06.653 15:23:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:18:06.653 15:23:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:06.653 15:23:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:18:06.653 15:23:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:06.913 15:23:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:18:06.913 "name": "raid_bdev1", 00:18:06.913 "uuid": "e025bbfe-cc1c-4ee3-97ed-a695e0cd3e6d", 00:18:06.913 "strip_size_kb": 0, 00:18:06.913 "state": "online", 00:18:06.913 "raid_level": "raid1", 00:18:06.913 "superblock": true, 00:18:06.913 "num_base_bdevs": 2, 00:18:06.913 "num_base_bdevs_discovered": 1, 00:18:06.913 "num_base_bdevs_operational": 1, 00:18:06.913 "base_bdevs_list": [ 00:18:06.913 { 00:18:06.913 "name": null, 00:18:06.913 "uuid": "00000000-0000-0000-0000-000000000000", 00:18:06.913 "is_configured": false, 00:18:06.913 "data_offset": 0, 00:18:06.913 "data_size": 7936 00:18:06.913 }, 00:18:06.913 { 00:18:06.913 "name": "BaseBdev2", 00:18:06.913 "uuid": "0bb42b7d-5683-528d-bf16-2cf8c9f3bdc3", 00:18:06.913 "is_configured": true, 00:18:06.913 "data_offset": 256, 00:18:06.913 "data_size": 7936 00:18:06.913 } 00:18:06.913 ] 00:18:06.913 }' 00:18:06.913 15:23:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:18:06.913 15:23:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:18:07.173 15:23:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:18:07.173 15:23:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:18:07.173 15:23:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:18:07.173 15:23:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:18:07.173 15:23:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:18:07.173 15:23:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:18:07.173 15:23:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:18:07.173 15:23:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:07.173 15:23:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:18:07.173 15:23:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:07.173 15:23:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:18:07.173 "name": "raid_bdev1", 00:18:07.173 "uuid": "e025bbfe-cc1c-4ee3-97ed-a695e0cd3e6d", 00:18:07.173 "strip_size_kb": 0, 00:18:07.173 "state": "online", 00:18:07.173 "raid_level": "raid1", 00:18:07.173 "superblock": true, 00:18:07.173 "num_base_bdevs": 2, 00:18:07.173 "num_base_bdevs_discovered": 1, 00:18:07.173 "num_base_bdevs_operational": 1, 00:18:07.173 "base_bdevs_list": [ 00:18:07.173 { 00:18:07.173 "name": null, 00:18:07.173 "uuid": "00000000-0000-0000-0000-000000000000", 00:18:07.173 "is_configured": false, 00:18:07.173 "data_offset": 0, 00:18:07.173 "data_size": 7936 00:18:07.174 }, 00:18:07.174 { 00:18:07.174 "name": "BaseBdev2", 00:18:07.174 "uuid": "0bb42b7d-5683-528d-bf16-2cf8c9f3bdc3", 00:18:07.174 "is_configured": true, 00:18:07.174 "data_offset": 256, 00:18:07.174 "data_size": 7936 00:18:07.174 } 00:18:07.174 ] 00:18:07.174 }' 00:18:07.174 15:23:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:18:07.174 15:23:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:18:07.174 15:23:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:18:07.434 15:23:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:18:07.434 15:23:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@784 -- # killprocess 99348 00:18:07.434 15:23:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@954 -- # '[' -z 99348 ']' 00:18:07.434 15:23:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@958 -- # kill -0 99348 00:18:07.434 15:23:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@959 -- # uname 00:18:07.434 15:23:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:18:07.434 15:23:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 99348 00:18:07.434 15:23:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:18:07.434 15:23:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:18:07.434 15:23:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@972 -- # echo 'killing process with pid 99348' 00:18:07.434 killing process with pid 99348 00:18:07.434 Received shutdown signal, test time was about 60.000000 seconds 00:18:07.434 00:18:07.434 Latency(us) 00:18:07.434 [2024-11-19T15:23:57.773Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:18:07.434 [2024-11-19T15:23:57.773Z] =================================================================================================================== 00:18:07.434 [2024-11-19T15:23:57.773Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:18:07.434 15:23:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@973 -- # kill 99348 00:18:07.434 [2024-11-19 15:23:57.567506] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:18:07.434 [2024-11-19 15:23:57.567623] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:18:07.434 15:23:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@978 -- # wait 99348 00:18:07.434 [2024-11-19 15:23:57.567669] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:18:07.434 [2024-11-19 15:23:57.567680] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:18:07.434 [2024-11-19 15:23:57.600598] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:18:07.695 15:23:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@786 -- # return 0 00:18:07.695 00:18:07.695 real 0m16.134s 00:18:07.695 user 0m21.576s 00:18:07.695 sys 0m1.699s 00:18:07.695 15:23:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@1130 -- # xtrace_disable 00:18:07.695 ************************************ 00:18:07.695 END TEST raid_rebuild_test_sb_md_interleaved 00:18:07.695 ************************************ 00:18:07.695 15:23:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:18:07.695 15:23:57 bdev_raid -- bdev/bdev_raid.sh@1015 -- # trap - EXIT 00:18:07.695 15:23:57 bdev_raid -- bdev/bdev_raid.sh@1016 -- # cleanup 00:18:07.695 15:23:57 bdev_raid -- bdev/bdev_raid.sh@56 -- # '[' -n 99348 ']' 00:18:07.695 15:23:57 bdev_raid -- bdev/bdev_raid.sh@56 -- # ps -p 99348 00:18:07.695 15:23:57 bdev_raid -- bdev/bdev_raid.sh@60 -- # rm -rf /raidtest 00:18:07.695 00:18:07.695 real 10m5.126s 00:18:07.695 user 14m12.383s 00:18:07.695 sys 1m54.041s 00:18:07.695 15:23:57 bdev_raid -- common/autotest_common.sh@1130 -- # xtrace_disable 00:18:07.695 15:23:57 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:18:07.695 ************************************ 00:18:07.695 END TEST bdev_raid 00:18:07.695 ************************************ 00:18:07.695 15:23:57 -- spdk/autotest.sh@190 -- # run_test spdkcli_raid /home/vagrant/spdk_repo/spdk/test/spdkcli/raid.sh 00:18:07.695 15:23:57 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:18:07.695 15:23:57 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:18:07.695 15:23:57 -- common/autotest_common.sh@10 -- # set +x 00:18:07.695 ************************************ 00:18:07.695 START TEST spdkcli_raid 00:18:07.695 ************************************ 00:18:07.695 15:23:57 spdkcli_raid -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/raid.sh 00:18:07.956 * Looking for test storage... 00:18:07.956 * Found test storage at /home/vagrant/spdk_repo/spdk/test/spdkcli 00:18:07.956 15:23:58 spdkcli_raid -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:18:07.956 15:23:58 spdkcli_raid -- common/autotest_common.sh@1693 -- # lcov --version 00:18:07.956 15:23:58 spdkcli_raid -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:18:07.956 15:23:58 spdkcli_raid -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:18:07.956 15:23:58 spdkcli_raid -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:18:07.956 15:23:58 spdkcli_raid -- scripts/common.sh@333 -- # local ver1 ver1_l 00:18:07.956 15:23:58 spdkcli_raid -- scripts/common.sh@334 -- # local ver2 ver2_l 00:18:07.956 15:23:58 spdkcli_raid -- scripts/common.sh@336 -- # IFS=.-: 00:18:07.956 15:23:58 spdkcli_raid -- scripts/common.sh@336 -- # read -ra ver1 00:18:07.956 15:23:58 spdkcli_raid -- scripts/common.sh@337 -- # IFS=.-: 00:18:07.956 15:23:58 spdkcli_raid -- scripts/common.sh@337 -- # read -ra ver2 00:18:07.956 15:23:58 spdkcli_raid -- scripts/common.sh@338 -- # local 'op=<' 00:18:07.956 15:23:58 spdkcli_raid -- scripts/common.sh@340 -- # ver1_l=2 00:18:07.956 15:23:58 spdkcli_raid -- scripts/common.sh@341 -- # ver2_l=1 00:18:07.956 15:23:58 spdkcli_raid -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:18:07.956 15:23:58 spdkcli_raid -- scripts/common.sh@344 -- # case "$op" in 00:18:07.956 15:23:58 spdkcli_raid -- scripts/common.sh@345 -- # : 1 00:18:07.956 15:23:58 spdkcli_raid -- scripts/common.sh@364 -- # (( v = 0 )) 00:18:07.956 15:23:58 spdkcli_raid -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:18:07.956 15:23:58 spdkcli_raid -- scripts/common.sh@365 -- # decimal 1 00:18:07.956 15:23:58 spdkcli_raid -- scripts/common.sh@353 -- # local d=1 00:18:07.956 15:23:58 spdkcli_raid -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:18:07.956 15:23:58 spdkcli_raid -- scripts/common.sh@355 -- # echo 1 00:18:07.956 15:23:58 spdkcli_raid -- scripts/common.sh@365 -- # ver1[v]=1 00:18:07.956 15:23:58 spdkcli_raid -- scripts/common.sh@366 -- # decimal 2 00:18:07.956 15:23:58 spdkcli_raid -- scripts/common.sh@353 -- # local d=2 00:18:07.956 15:23:58 spdkcli_raid -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:18:07.956 15:23:58 spdkcli_raid -- scripts/common.sh@355 -- # echo 2 00:18:07.956 15:23:58 spdkcli_raid -- scripts/common.sh@366 -- # ver2[v]=2 00:18:07.956 15:23:58 spdkcli_raid -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:18:07.956 15:23:58 spdkcli_raid -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:18:07.956 15:23:58 spdkcli_raid -- scripts/common.sh@368 -- # return 0 00:18:07.956 15:23:58 spdkcli_raid -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:18:07.956 15:23:58 spdkcli_raid -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:18:07.956 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:07.956 --rc genhtml_branch_coverage=1 00:18:07.956 --rc genhtml_function_coverage=1 00:18:07.956 --rc genhtml_legend=1 00:18:07.956 --rc geninfo_all_blocks=1 00:18:07.956 --rc geninfo_unexecuted_blocks=1 00:18:07.956 00:18:07.956 ' 00:18:07.956 15:23:58 spdkcli_raid -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:18:07.956 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:07.956 --rc genhtml_branch_coverage=1 00:18:07.956 --rc genhtml_function_coverage=1 00:18:07.956 --rc genhtml_legend=1 00:18:07.956 --rc geninfo_all_blocks=1 00:18:07.956 --rc geninfo_unexecuted_blocks=1 00:18:07.956 00:18:07.956 ' 00:18:07.956 15:23:58 spdkcli_raid -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:18:07.956 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:07.956 --rc genhtml_branch_coverage=1 00:18:07.956 --rc genhtml_function_coverage=1 00:18:07.956 --rc genhtml_legend=1 00:18:07.956 --rc geninfo_all_blocks=1 00:18:07.956 --rc geninfo_unexecuted_blocks=1 00:18:07.956 00:18:07.956 ' 00:18:07.956 15:23:58 spdkcli_raid -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:18:07.956 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:07.956 --rc genhtml_branch_coverage=1 00:18:07.956 --rc genhtml_function_coverage=1 00:18:07.956 --rc genhtml_legend=1 00:18:07.956 --rc geninfo_all_blocks=1 00:18:07.956 --rc geninfo_unexecuted_blocks=1 00:18:07.956 00:18:07.956 ' 00:18:07.956 15:23:58 spdkcli_raid -- spdkcli/raid.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/spdkcli/common.sh 00:18:07.956 15:23:58 spdkcli_raid -- spdkcli/common.sh@6 -- # spdkcli_job=/home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py 00:18:07.956 15:23:58 spdkcli_raid -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/home/vagrant/spdk_repo/spdk/test/json_config/clear_config.py 00:18:07.956 15:23:58 spdkcli_raid -- spdkcli/raid.sh@10 -- # source /home/vagrant/spdk_repo/spdk/test/iscsi_tgt/common.sh 00:18:07.956 15:23:58 spdkcli_raid -- iscsi_tgt/common.sh@9 -- # ISCSI_BRIDGE=iscsi_br 00:18:07.956 15:23:58 spdkcli_raid -- iscsi_tgt/common.sh@10 -- # INITIATOR_INTERFACE=spdk_init_int 00:18:07.956 15:23:58 spdkcli_raid -- iscsi_tgt/common.sh@11 -- # INITIATOR_BRIDGE=init_br 00:18:07.956 15:23:58 spdkcli_raid -- iscsi_tgt/common.sh@12 -- # TARGET_NAMESPACE=spdk_iscsi_ns 00:18:07.956 15:23:58 spdkcli_raid -- iscsi_tgt/common.sh@13 -- # TARGET_NS_CMD=(ip netns exec "$TARGET_NAMESPACE") 00:18:07.956 15:23:58 spdkcli_raid -- iscsi_tgt/common.sh@14 -- # TARGET_INTERFACE=spdk_tgt_int 00:18:07.956 15:23:58 spdkcli_raid -- iscsi_tgt/common.sh@15 -- # TARGET_INTERFACE2=spdk_tgt_int2 00:18:07.956 15:23:58 spdkcli_raid -- iscsi_tgt/common.sh@16 -- # TARGET_BRIDGE=tgt_br 00:18:07.956 15:23:58 spdkcli_raid -- iscsi_tgt/common.sh@17 -- # TARGET_BRIDGE2=tgt_br2 00:18:07.956 15:23:58 spdkcli_raid -- iscsi_tgt/common.sh@20 -- # TARGET_IP=10.0.0.1 00:18:07.956 15:23:58 spdkcli_raid -- iscsi_tgt/common.sh@21 -- # TARGET_IP2=10.0.0.3 00:18:07.956 15:23:58 spdkcli_raid -- iscsi_tgt/common.sh@22 -- # INITIATOR_IP=10.0.0.2 00:18:07.956 15:23:58 spdkcli_raid -- iscsi_tgt/common.sh@23 -- # ISCSI_PORT=3260 00:18:07.956 15:23:58 spdkcli_raid -- iscsi_tgt/common.sh@24 -- # NETMASK=10.0.0.2/32 00:18:07.956 15:23:58 spdkcli_raid -- iscsi_tgt/common.sh@25 -- # INITIATOR_TAG=2 00:18:07.956 15:23:58 spdkcli_raid -- iscsi_tgt/common.sh@26 -- # INITIATOR_NAME=ANY 00:18:07.956 15:23:58 spdkcli_raid -- iscsi_tgt/common.sh@27 -- # PORTAL_TAG=1 00:18:07.956 15:23:58 spdkcli_raid -- iscsi_tgt/common.sh@28 -- # ISCSI_APP=("${TARGET_NS_CMD[@]}" "${ISCSI_APP[@]}") 00:18:07.956 15:23:58 spdkcli_raid -- iscsi_tgt/common.sh@29 -- # ISCSI_TEST_CORE_MASK=0xF 00:18:07.956 15:23:58 spdkcli_raid -- spdkcli/raid.sh@12 -- # MATCH_FILE=spdkcli_raid.test 00:18:07.956 15:23:58 spdkcli_raid -- spdkcli/raid.sh@13 -- # SPDKCLI_BRANCH=/bdevs 00:18:07.956 15:23:58 spdkcli_raid -- spdkcli/raid.sh@14 -- # dirname /home/vagrant/spdk_repo/spdk/test/spdkcli/raid.sh 00:18:07.956 15:23:58 spdkcli_raid -- spdkcli/raid.sh@14 -- # readlink -f /home/vagrant/spdk_repo/spdk/test/spdkcli 00:18:07.956 15:23:58 spdkcli_raid -- spdkcli/raid.sh@14 -- # testdir=/home/vagrant/spdk_repo/spdk/test/spdkcli 00:18:07.956 15:23:58 spdkcli_raid -- spdkcli/raid.sh@15 -- # . /home/vagrant/spdk_repo/spdk/test/spdkcli/common.sh 00:18:07.956 15:23:58 spdkcli_raid -- spdkcli/common.sh@6 -- # spdkcli_job=/home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py 00:18:07.956 15:23:58 spdkcli_raid -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/home/vagrant/spdk_repo/spdk/test/json_config/clear_config.py 00:18:07.956 15:23:58 spdkcli_raid -- spdkcli/raid.sh@17 -- # trap cleanup EXIT 00:18:07.956 15:23:58 spdkcli_raid -- spdkcli/raid.sh@19 -- # timing_enter run_spdk_tgt 00:18:07.956 15:23:58 spdkcli_raid -- common/autotest_common.sh@726 -- # xtrace_disable 00:18:07.956 15:23:58 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:18:07.956 15:23:58 spdkcli_raid -- spdkcli/raid.sh@20 -- # run_spdk_tgt 00:18:07.956 15:23:58 spdkcli_raid -- spdkcli/common.sh@27 -- # spdk_tgt_pid=100012 00:18:07.956 15:23:58 spdkcli_raid -- spdkcli/common.sh@26 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x3 -p 0 00:18:07.956 15:23:58 spdkcli_raid -- spdkcli/common.sh@28 -- # waitforlisten 100012 00:18:07.956 15:23:58 spdkcli_raid -- common/autotest_common.sh@835 -- # '[' -z 100012 ']' 00:18:07.956 15:23:58 spdkcli_raid -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:18:07.956 15:23:58 spdkcli_raid -- common/autotest_common.sh@840 -- # local max_retries=100 00:18:07.956 15:23:58 spdkcli_raid -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:18:07.956 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:18:07.956 15:23:58 spdkcli_raid -- common/autotest_common.sh@844 -- # xtrace_disable 00:18:07.956 15:23:58 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:18:08.216 [2024-11-19 15:23:58.335954] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:18:08.216 [2024-11-19 15:23:58.336177] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100012 ] 00:18:08.217 [2024-11-19 15:23:58.493334] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:18:08.217 [2024-11-19 15:23:58.520908] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:18:08.217 [2024-11-19 15:23:58.521032] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:18:08.786 15:23:59 spdkcli_raid -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:18:08.786 15:23:59 spdkcli_raid -- common/autotest_common.sh@868 -- # return 0 00:18:08.786 15:23:59 spdkcli_raid -- spdkcli/raid.sh@21 -- # timing_exit run_spdk_tgt 00:18:08.786 15:23:59 spdkcli_raid -- common/autotest_common.sh@732 -- # xtrace_disable 00:18:08.786 15:23:59 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:18:09.045 15:23:59 spdkcli_raid -- spdkcli/raid.sh@23 -- # timing_enter spdkcli_create_malloc 00:18:09.045 15:23:59 spdkcli_raid -- common/autotest_common.sh@726 -- # xtrace_disable 00:18:09.045 15:23:59 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:18:09.045 15:23:59 spdkcli_raid -- spdkcli/raid.sh@26 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/malloc create 8 512 Malloc1'\'' '\''Malloc1'\'' True 00:18:09.045 '\''/bdevs/malloc create 8 512 Malloc2'\'' '\''Malloc2'\'' True 00:18:09.045 ' 00:18:10.428 Executing command: ['/bdevs/malloc create 8 512 Malloc1', 'Malloc1', True] 00:18:10.428 Executing command: ['/bdevs/malloc create 8 512 Malloc2', 'Malloc2', True] 00:18:10.688 15:24:00 spdkcli_raid -- spdkcli/raid.sh@27 -- # timing_exit spdkcli_create_malloc 00:18:10.688 15:24:00 spdkcli_raid -- common/autotest_common.sh@732 -- # xtrace_disable 00:18:10.688 15:24:00 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:18:10.688 15:24:00 spdkcli_raid -- spdkcli/raid.sh@29 -- # timing_enter spdkcli_create_raid 00:18:10.688 15:24:00 spdkcli_raid -- common/autotest_common.sh@726 -- # xtrace_disable 00:18:10.688 15:24:00 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:18:10.689 15:24:00 spdkcli_raid -- spdkcli/raid.sh@31 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/raid_volume create testraid 0 "Malloc1 Malloc2" 4'\'' '\''testraid'\'' True 00:18:10.689 ' 00:18:11.628 Executing command: ['/bdevs/raid_volume create testraid 0 "Malloc1 Malloc2" 4', 'testraid', True] 00:18:11.887 15:24:01 spdkcli_raid -- spdkcli/raid.sh@32 -- # timing_exit spdkcli_create_raid 00:18:11.887 15:24:01 spdkcli_raid -- common/autotest_common.sh@732 -- # xtrace_disable 00:18:11.887 15:24:01 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:18:11.887 15:24:02 spdkcli_raid -- spdkcli/raid.sh@34 -- # timing_enter spdkcli_check_match 00:18:11.887 15:24:02 spdkcli_raid -- common/autotest_common.sh@726 -- # xtrace_disable 00:18:11.887 15:24:02 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:18:11.887 15:24:02 spdkcli_raid -- spdkcli/raid.sh@35 -- # check_match 00:18:11.887 15:24:02 spdkcli_raid -- spdkcli/common.sh@44 -- # /home/vagrant/spdk_repo/spdk/scripts/spdkcli.py ll /bdevs 00:18:12.456 15:24:02 spdkcli_raid -- spdkcli/common.sh@45 -- # /home/vagrant/spdk_repo/spdk/test/app/match/match /home/vagrant/spdk_repo/spdk/test/spdkcli/match_files/spdkcli_raid.test.match 00:18:12.456 15:24:02 spdkcli_raid -- spdkcli/common.sh@46 -- # rm -f /home/vagrant/spdk_repo/spdk/test/spdkcli/match_files/spdkcli_raid.test 00:18:12.456 15:24:02 spdkcli_raid -- spdkcli/raid.sh@36 -- # timing_exit spdkcli_check_match 00:18:12.456 15:24:02 spdkcli_raid -- common/autotest_common.sh@732 -- # xtrace_disable 00:18:12.456 15:24:02 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:18:12.456 15:24:02 spdkcli_raid -- spdkcli/raid.sh@38 -- # timing_enter spdkcli_delete_raid 00:18:12.456 15:24:02 spdkcli_raid -- common/autotest_common.sh@726 -- # xtrace_disable 00:18:12.456 15:24:02 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:18:12.456 15:24:02 spdkcli_raid -- spdkcli/raid.sh@40 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/raid_volume delete testraid'\'' '\'''\'' True 00:18:12.456 ' 00:18:13.395 Executing command: ['/bdevs/raid_volume delete testraid', '', True] 00:18:13.395 15:24:03 spdkcli_raid -- spdkcli/raid.sh@41 -- # timing_exit spdkcli_delete_raid 00:18:13.395 15:24:03 spdkcli_raid -- common/autotest_common.sh@732 -- # xtrace_disable 00:18:13.395 15:24:03 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:18:13.655 15:24:03 spdkcli_raid -- spdkcli/raid.sh@43 -- # timing_enter spdkcli_delete_malloc 00:18:13.655 15:24:03 spdkcli_raid -- common/autotest_common.sh@726 -- # xtrace_disable 00:18:13.655 15:24:03 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:18:13.655 15:24:03 spdkcli_raid -- spdkcli/raid.sh@46 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/malloc delete Malloc1'\'' '\'''\'' True 00:18:13.655 '\''/bdevs/malloc delete Malloc2'\'' '\'''\'' True 00:18:13.655 ' 00:18:15.038 Executing command: ['/bdevs/malloc delete Malloc1', '', True] 00:18:15.038 Executing command: ['/bdevs/malloc delete Malloc2', '', True] 00:18:15.038 15:24:05 spdkcli_raid -- spdkcli/raid.sh@47 -- # timing_exit spdkcli_delete_malloc 00:18:15.038 15:24:05 spdkcli_raid -- common/autotest_common.sh@732 -- # xtrace_disable 00:18:15.038 15:24:05 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:18:15.038 15:24:05 spdkcli_raid -- spdkcli/raid.sh@49 -- # killprocess 100012 00:18:15.038 15:24:05 spdkcli_raid -- common/autotest_common.sh@954 -- # '[' -z 100012 ']' 00:18:15.038 15:24:05 spdkcli_raid -- common/autotest_common.sh@958 -- # kill -0 100012 00:18:15.038 15:24:05 spdkcli_raid -- common/autotest_common.sh@959 -- # uname 00:18:15.038 15:24:05 spdkcli_raid -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:18:15.038 15:24:05 spdkcli_raid -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 100012 00:18:15.038 15:24:05 spdkcli_raid -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:18:15.038 15:24:05 spdkcli_raid -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:18:15.038 15:24:05 spdkcli_raid -- common/autotest_common.sh@972 -- # echo 'killing process with pid 100012' 00:18:15.038 killing process with pid 100012 00:18:15.038 15:24:05 spdkcli_raid -- common/autotest_common.sh@973 -- # kill 100012 00:18:15.038 15:24:05 spdkcli_raid -- common/autotest_common.sh@978 -- # wait 100012 00:18:15.609 15:24:05 spdkcli_raid -- spdkcli/raid.sh@1 -- # cleanup 00:18:15.609 Process with pid 100012 is not found 00:18:15.609 15:24:05 spdkcli_raid -- spdkcli/common.sh@10 -- # '[' -n 100012 ']' 00:18:15.609 15:24:05 spdkcli_raid -- spdkcli/common.sh@11 -- # killprocess 100012 00:18:15.609 15:24:05 spdkcli_raid -- common/autotest_common.sh@954 -- # '[' -z 100012 ']' 00:18:15.609 15:24:05 spdkcli_raid -- common/autotest_common.sh@958 -- # kill -0 100012 00:18:15.609 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 958: kill: (100012) - No such process 00:18:15.609 15:24:05 spdkcli_raid -- common/autotest_common.sh@981 -- # echo 'Process with pid 100012 is not found' 00:18:15.609 15:24:05 spdkcli_raid -- spdkcli/common.sh@13 -- # '[' -n '' ']' 00:18:15.609 15:24:05 spdkcli_raid -- spdkcli/common.sh@16 -- # '[' -n '' ']' 00:18:15.609 15:24:05 spdkcli_raid -- spdkcli/common.sh@19 -- # '[' -n '' ']' 00:18:15.609 15:24:05 spdkcli_raid -- spdkcli/common.sh@22 -- # rm -f /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_raid.test /home/vagrant/spdk_repo/spdk/test/spdkcli/match_files/spdkcli_details_vhost.test /tmp/sample_aio 00:18:15.609 00:18:15.609 real 0m7.705s 00:18:15.609 user 0m16.207s 00:18:15.609 sys 0m1.135s 00:18:15.609 15:24:05 spdkcli_raid -- common/autotest_common.sh@1130 -- # xtrace_disable 00:18:15.609 15:24:05 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:18:15.609 ************************************ 00:18:15.609 END TEST spdkcli_raid 00:18:15.609 ************************************ 00:18:15.609 15:24:05 -- spdk/autotest.sh@191 -- # run_test blockdev_raid5f /home/vagrant/spdk_repo/spdk/test/bdev/blockdev.sh raid5f 00:18:15.609 15:24:05 -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:18:15.609 15:24:05 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:18:15.609 15:24:05 -- common/autotest_common.sh@10 -- # set +x 00:18:15.609 ************************************ 00:18:15.609 START TEST blockdev_raid5f 00:18:15.609 ************************************ 00:18:15.609 15:24:05 blockdev_raid5f -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/bdev/blockdev.sh raid5f 00:18:15.609 * Looking for test storage... 00:18:15.609 * Found test storage at /home/vagrant/spdk_repo/spdk/test/bdev 00:18:15.609 15:24:05 blockdev_raid5f -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:18:15.609 15:24:05 blockdev_raid5f -- common/autotest_common.sh@1693 -- # lcov --version 00:18:15.609 15:24:05 blockdev_raid5f -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:18:15.868 15:24:05 blockdev_raid5f -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:18:15.869 15:24:05 blockdev_raid5f -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:18:15.869 15:24:05 blockdev_raid5f -- scripts/common.sh@333 -- # local ver1 ver1_l 00:18:15.869 15:24:05 blockdev_raid5f -- scripts/common.sh@334 -- # local ver2 ver2_l 00:18:15.869 15:24:05 blockdev_raid5f -- scripts/common.sh@336 -- # IFS=.-: 00:18:15.869 15:24:05 blockdev_raid5f -- scripts/common.sh@336 -- # read -ra ver1 00:18:15.869 15:24:05 blockdev_raid5f -- scripts/common.sh@337 -- # IFS=.-: 00:18:15.869 15:24:05 blockdev_raid5f -- scripts/common.sh@337 -- # read -ra ver2 00:18:15.869 15:24:05 blockdev_raid5f -- scripts/common.sh@338 -- # local 'op=<' 00:18:15.869 15:24:05 blockdev_raid5f -- scripts/common.sh@340 -- # ver1_l=2 00:18:15.869 15:24:05 blockdev_raid5f -- scripts/common.sh@341 -- # ver2_l=1 00:18:15.869 15:24:05 blockdev_raid5f -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:18:15.869 15:24:05 blockdev_raid5f -- scripts/common.sh@344 -- # case "$op" in 00:18:15.869 15:24:05 blockdev_raid5f -- scripts/common.sh@345 -- # : 1 00:18:15.869 15:24:05 blockdev_raid5f -- scripts/common.sh@364 -- # (( v = 0 )) 00:18:15.869 15:24:05 blockdev_raid5f -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:18:15.869 15:24:05 blockdev_raid5f -- scripts/common.sh@365 -- # decimal 1 00:18:15.869 15:24:05 blockdev_raid5f -- scripts/common.sh@353 -- # local d=1 00:18:15.869 15:24:05 blockdev_raid5f -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:18:15.869 15:24:05 blockdev_raid5f -- scripts/common.sh@355 -- # echo 1 00:18:15.869 15:24:05 blockdev_raid5f -- scripts/common.sh@365 -- # ver1[v]=1 00:18:15.869 15:24:05 blockdev_raid5f -- scripts/common.sh@366 -- # decimal 2 00:18:15.869 15:24:05 blockdev_raid5f -- scripts/common.sh@353 -- # local d=2 00:18:15.869 15:24:05 blockdev_raid5f -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:18:15.869 15:24:05 blockdev_raid5f -- scripts/common.sh@355 -- # echo 2 00:18:15.869 15:24:05 blockdev_raid5f -- scripts/common.sh@366 -- # ver2[v]=2 00:18:15.869 15:24:05 blockdev_raid5f -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:18:15.869 15:24:05 blockdev_raid5f -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:18:15.869 15:24:05 blockdev_raid5f -- scripts/common.sh@368 -- # return 0 00:18:15.869 15:24:05 blockdev_raid5f -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:18:15.869 15:24:05 blockdev_raid5f -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:18:15.869 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:15.869 --rc genhtml_branch_coverage=1 00:18:15.869 --rc genhtml_function_coverage=1 00:18:15.869 --rc genhtml_legend=1 00:18:15.869 --rc geninfo_all_blocks=1 00:18:15.869 --rc geninfo_unexecuted_blocks=1 00:18:15.869 00:18:15.869 ' 00:18:15.869 15:24:05 blockdev_raid5f -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:18:15.869 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:15.869 --rc genhtml_branch_coverage=1 00:18:15.869 --rc genhtml_function_coverage=1 00:18:15.869 --rc genhtml_legend=1 00:18:15.869 --rc geninfo_all_blocks=1 00:18:15.869 --rc geninfo_unexecuted_blocks=1 00:18:15.869 00:18:15.869 ' 00:18:15.869 15:24:05 blockdev_raid5f -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:18:15.869 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:15.869 --rc genhtml_branch_coverage=1 00:18:15.869 --rc genhtml_function_coverage=1 00:18:15.869 --rc genhtml_legend=1 00:18:15.869 --rc geninfo_all_blocks=1 00:18:15.869 --rc geninfo_unexecuted_blocks=1 00:18:15.869 00:18:15.869 ' 00:18:15.869 15:24:05 blockdev_raid5f -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:18:15.869 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:15.869 --rc genhtml_branch_coverage=1 00:18:15.869 --rc genhtml_function_coverage=1 00:18:15.869 --rc genhtml_legend=1 00:18:15.869 --rc geninfo_all_blocks=1 00:18:15.869 --rc geninfo_unexecuted_blocks=1 00:18:15.869 00:18:15.869 ' 00:18:15.869 15:24:05 blockdev_raid5f -- bdev/blockdev.sh@10 -- # source /home/vagrant/spdk_repo/spdk/test/bdev/nbd_common.sh 00:18:15.869 15:24:05 blockdev_raid5f -- bdev/nbd_common.sh@6 -- # set -e 00:18:15.869 15:24:05 blockdev_raid5f -- bdev/blockdev.sh@12 -- # rpc_py=rpc_cmd 00:18:15.869 15:24:05 blockdev_raid5f -- bdev/blockdev.sh@13 -- # conf_file=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json 00:18:15.869 15:24:05 blockdev_raid5f -- bdev/blockdev.sh@14 -- # nonenclosed_conf_file=/home/vagrant/spdk_repo/spdk/test/bdev/nonenclosed.json 00:18:15.869 15:24:05 blockdev_raid5f -- bdev/blockdev.sh@15 -- # nonarray_conf_file=/home/vagrant/spdk_repo/spdk/test/bdev/nonarray.json 00:18:15.869 15:24:05 blockdev_raid5f -- bdev/blockdev.sh@17 -- # export RPC_PIPE_TIMEOUT=30 00:18:15.869 15:24:05 blockdev_raid5f -- bdev/blockdev.sh@17 -- # RPC_PIPE_TIMEOUT=30 00:18:15.869 15:24:05 blockdev_raid5f -- bdev/blockdev.sh@20 -- # : 00:18:15.869 15:24:05 blockdev_raid5f -- bdev/blockdev.sh@669 -- # QOS_DEV_1=Malloc_0 00:18:15.869 15:24:05 blockdev_raid5f -- bdev/blockdev.sh@670 -- # QOS_DEV_2=Null_1 00:18:15.869 15:24:05 blockdev_raid5f -- bdev/blockdev.sh@671 -- # QOS_RUN_TIME=5 00:18:15.869 15:24:05 blockdev_raid5f -- bdev/blockdev.sh@673 -- # uname -s 00:18:15.869 15:24:05 blockdev_raid5f -- bdev/blockdev.sh@673 -- # '[' Linux = Linux ']' 00:18:15.869 15:24:05 blockdev_raid5f -- bdev/blockdev.sh@675 -- # PRE_RESERVED_MEM=0 00:18:15.869 15:24:05 blockdev_raid5f -- bdev/blockdev.sh@681 -- # test_type=raid5f 00:18:15.869 15:24:05 blockdev_raid5f -- bdev/blockdev.sh@682 -- # crypto_device= 00:18:15.869 15:24:05 blockdev_raid5f -- bdev/blockdev.sh@683 -- # dek= 00:18:15.869 15:24:05 blockdev_raid5f -- bdev/blockdev.sh@684 -- # env_ctx= 00:18:15.869 15:24:05 blockdev_raid5f -- bdev/blockdev.sh@685 -- # wait_for_rpc= 00:18:15.869 15:24:05 blockdev_raid5f -- bdev/blockdev.sh@686 -- # '[' -n '' ']' 00:18:15.869 15:24:05 blockdev_raid5f -- bdev/blockdev.sh@689 -- # [[ raid5f == bdev ]] 00:18:15.869 15:24:05 blockdev_raid5f -- bdev/blockdev.sh@689 -- # [[ raid5f == crypto_* ]] 00:18:15.869 15:24:05 blockdev_raid5f -- bdev/blockdev.sh@692 -- # start_spdk_tgt 00:18:15.869 15:24:06 blockdev_raid5f -- bdev/blockdev.sh@47 -- # spdk_tgt_pid=100264 00:18:15.869 15:24:06 blockdev_raid5f -- bdev/blockdev.sh@46 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt '' '' 00:18:15.869 15:24:06 blockdev_raid5f -- bdev/blockdev.sh@48 -- # trap 'killprocess "$spdk_tgt_pid"; exit 1' SIGINT SIGTERM EXIT 00:18:15.869 15:24:06 blockdev_raid5f -- bdev/blockdev.sh@49 -- # waitforlisten 100264 00:18:15.869 15:24:06 blockdev_raid5f -- common/autotest_common.sh@835 -- # '[' -z 100264 ']' 00:18:15.869 15:24:06 blockdev_raid5f -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:18:15.869 15:24:06 blockdev_raid5f -- common/autotest_common.sh@840 -- # local max_retries=100 00:18:15.869 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:18:15.869 15:24:06 blockdev_raid5f -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:18:15.869 15:24:06 blockdev_raid5f -- common/autotest_common.sh@844 -- # xtrace_disable 00:18:15.869 15:24:06 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:18:15.869 [2024-11-19 15:24:06.093903] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:18:15.869 [2024-11-19 15:24:06.094045] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100264 ] 00:18:16.128 [2024-11-19 15:24:06.247778] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:18:16.128 [2024-11-19 15:24:06.273727] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:18:16.698 15:24:06 blockdev_raid5f -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:18:16.698 15:24:06 blockdev_raid5f -- common/autotest_common.sh@868 -- # return 0 00:18:16.698 15:24:06 blockdev_raid5f -- bdev/blockdev.sh@693 -- # case "$test_type" in 00:18:16.698 15:24:06 blockdev_raid5f -- bdev/blockdev.sh@725 -- # setup_raid5f_conf 00:18:16.698 15:24:06 blockdev_raid5f -- bdev/blockdev.sh@279 -- # rpc_cmd 00:18:16.698 15:24:06 blockdev_raid5f -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:16.698 15:24:06 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:18:16.699 Malloc0 00:18:16.699 Malloc1 00:18:16.699 Malloc2 00:18:16.699 15:24:06 blockdev_raid5f -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:16.699 15:24:06 blockdev_raid5f -- bdev/blockdev.sh@736 -- # rpc_cmd bdev_wait_for_examine 00:18:16.699 15:24:06 blockdev_raid5f -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:16.699 15:24:06 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:18:16.699 15:24:06 blockdev_raid5f -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:16.699 15:24:06 blockdev_raid5f -- bdev/blockdev.sh@739 -- # cat 00:18:16.699 15:24:06 blockdev_raid5f -- bdev/blockdev.sh@739 -- # rpc_cmd save_subsystem_config -n accel 00:18:16.699 15:24:06 blockdev_raid5f -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:16.699 15:24:06 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:18:16.699 15:24:06 blockdev_raid5f -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:16.699 15:24:06 blockdev_raid5f -- bdev/blockdev.sh@739 -- # rpc_cmd save_subsystem_config -n bdev 00:18:16.699 15:24:06 blockdev_raid5f -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:16.699 15:24:06 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:18:16.699 15:24:07 blockdev_raid5f -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:16.699 15:24:07 blockdev_raid5f -- bdev/blockdev.sh@739 -- # rpc_cmd save_subsystem_config -n iobuf 00:18:16.699 15:24:07 blockdev_raid5f -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:16.699 15:24:07 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:18:16.958 15:24:07 blockdev_raid5f -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:16.958 15:24:07 blockdev_raid5f -- bdev/blockdev.sh@747 -- # mapfile -t bdevs 00:18:16.958 15:24:07 blockdev_raid5f -- bdev/blockdev.sh@747 -- # rpc_cmd bdev_get_bdevs 00:18:16.958 15:24:07 blockdev_raid5f -- common/autotest_common.sh@563 -- # xtrace_disable 00:18:16.959 15:24:07 blockdev_raid5f -- bdev/blockdev.sh@747 -- # jq -r '.[] | select(.claimed == false)' 00:18:16.959 15:24:07 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:18:16.959 15:24:07 blockdev_raid5f -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:18:16.959 15:24:07 blockdev_raid5f -- bdev/blockdev.sh@748 -- # mapfile -t bdevs_name 00:18:16.959 15:24:07 blockdev_raid5f -- bdev/blockdev.sh@748 -- # printf '%s\n' '{' ' "name": "raid5f",' ' "aliases": [' ' "9420d3b3-3065-48e8-a5e5-7b2bc31c6b72"' ' ],' ' "product_name": "Raid Volume",' ' "block_size": 512,' ' "num_blocks": 131072,' ' "uuid": "9420d3b3-3065-48e8-a5e5-7b2bc31c6b72",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": false,' ' "flush": false,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "raid": {' ' "uuid": "9420d3b3-3065-48e8-a5e5-7b2bc31c6b72",' ' "strip_size_kb": 2,' ' "state": "online",' ' "raid_level": "raid5f",' ' "superblock": false,' ' "num_base_bdevs": 3,' ' "num_base_bdevs_discovered": 3,' ' "num_base_bdevs_operational": 3,' ' "base_bdevs_list": [' ' {' ' "name": "Malloc0",' ' "uuid": "f14beae4-660e-4035-9bac-d91415d8c490",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc1",' ' "uuid": "87a81309-869b-4719-b95c-29beb05627f6",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc2",' ' "uuid": "a835ec76-681d-46c6-acec-65e504f5880d",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' }' ' ]' ' }' ' }' '}' 00:18:16.959 15:24:07 blockdev_raid5f -- bdev/blockdev.sh@748 -- # jq -r .name 00:18:16.959 15:24:07 blockdev_raid5f -- bdev/blockdev.sh@749 -- # bdev_list=("${bdevs_name[@]}") 00:18:16.959 15:24:07 blockdev_raid5f -- bdev/blockdev.sh@751 -- # hello_world_bdev=raid5f 00:18:16.959 15:24:07 blockdev_raid5f -- bdev/blockdev.sh@752 -- # trap - SIGINT SIGTERM EXIT 00:18:16.959 15:24:07 blockdev_raid5f -- bdev/blockdev.sh@753 -- # killprocess 100264 00:18:16.959 15:24:07 blockdev_raid5f -- common/autotest_common.sh@954 -- # '[' -z 100264 ']' 00:18:16.959 15:24:07 blockdev_raid5f -- common/autotest_common.sh@958 -- # kill -0 100264 00:18:16.959 15:24:07 blockdev_raid5f -- common/autotest_common.sh@959 -- # uname 00:18:16.959 15:24:07 blockdev_raid5f -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:18:16.959 15:24:07 blockdev_raid5f -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 100264 00:18:16.959 killing process with pid 100264 00:18:16.959 15:24:07 blockdev_raid5f -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:18:16.959 15:24:07 blockdev_raid5f -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:18:16.959 15:24:07 blockdev_raid5f -- common/autotest_common.sh@972 -- # echo 'killing process with pid 100264' 00:18:16.959 15:24:07 blockdev_raid5f -- common/autotest_common.sh@973 -- # kill 100264 00:18:16.959 15:24:07 blockdev_raid5f -- common/autotest_common.sh@978 -- # wait 100264 00:18:17.528 15:24:07 blockdev_raid5f -- bdev/blockdev.sh@757 -- # trap cleanup SIGINT SIGTERM EXIT 00:18:17.528 15:24:07 blockdev_raid5f -- bdev/blockdev.sh@759 -- # run_test bdev_hello_world /home/vagrant/spdk_repo/spdk/build/examples/hello_bdev --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -b raid5f '' 00:18:17.529 15:24:07 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:18:17.529 15:24:07 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:18:17.529 15:24:07 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:18:17.529 ************************************ 00:18:17.529 START TEST bdev_hello_world 00:18:17.529 ************************************ 00:18:17.529 15:24:07 blockdev_raid5f.bdev_hello_world -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/build/examples/hello_bdev --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -b raid5f '' 00:18:17.529 [2024-11-19 15:24:07.682999] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:18:17.529 [2024-11-19 15:24:07.683189] [ DPDK EAL parameters: hello_bdev --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100304 ] 00:18:17.529 [2024-11-19 15:24:07.838737] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:18:17.529 [2024-11-19 15:24:07.864073] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:18:17.789 [2024-11-19 15:24:08.039366] hello_bdev.c: 222:hello_start: *NOTICE*: Successfully started the application 00:18:17.789 [2024-11-19 15:24:08.039414] hello_bdev.c: 231:hello_start: *NOTICE*: Opening the bdev raid5f 00:18:17.789 [2024-11-19 15:24:08.039430] hello_bdev.c: 244:hello_start: *NOTICE*: Opening io channel 00:18:17.789 [2024-11-19 15:24:08.039708] hello_bdev.c: 138:hello_write: *NOTICE*: Writing to the bdev 00:18:17.789 [2024-11-19 15:24:08.039850] hello_bdev.c: 117:write_complete: *NOTICE*: bdev io write completed successfully 00:18:17.789 [2024-11-19 15:24:08.039876] hello_bdev.c: 84:hello_read: *NOTICE*: Reading io 00:18:17.789 [2024-11-19 15:24:08.039920] hello_bdev.c: 65:read_complete: *NOTICE*: Read string from bdev : Hello World! 00:18:17.789 00:18:17.789 [2024-11-19 15:24:08.039936] hello_bdev.c: 74:read_complete: *NOTICE*: Stopping app 00:18:18.049 00:18:18.049 real 0m0.665s 00:18:18.049 user 0m0.366s 00:18:18.049 sys 0m0.193s 00:18:18.049 15:24:08 blockdev_raid5f.bdev_hello_world -- common/autotest_common.sh@1130 -- # xtrace_disable 00:18:18.049 ************************************ 00:18:18.049 END TEST bdev_hello_world 00:18:18.049 ************************************ 00:18:18.049 15:24:08 blockdev_raid5f.bdev_hello_world -- common/autotest_common.sh@10 -- # set +x 00:18:18.049 15:24:08 blockdev_raid5f -- bdev/blockdev.sh@760 -- # run_test bdev_bounds bdev_bounds '' 00:18:18.049 15:24:08 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:18:18.049 15:24:08 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:18:18.049 15:24:08 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:18:18.049 ************************************ 00:18:18.049 START TEST bdev_bounds 00:18:18.049 ************************************ 00:18:18.049 15:24:08 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@1129 -- # bdev_bounds '' 00:18:18.049 15:24:08 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@289 -- # bdevio_pid=100335 00:18:18.049 15:24:08 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@288 -- # /home/vagrant/spdk_repo/spdk/test/bdev/bdevio/bdevio -w -s 0 --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json '' 00:18:18.049 15:24:08 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@290 -- # trap 'cleanup; killprocess $bdevio_pid; exit 1' SIGINT SIGTERM EXIT 00:18:18.049 15:24:08 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@291 -- # echo 'Process bdevio pid: 100335' 00:18:18.049 Process bdevio pid: 100335 00:18:18.049 15:24:08 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@292 -- # waitforlisten 100335 00:18:18.049 15:24:08 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@835 -- # '[' -z 100335 ']' 00:18:18.049 15:24:08 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:18:18.049 15:24:08 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@840 -- # local max_retries=100 00:18:18.049 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:18:18.049 15:24:08 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:18:18.049 15:24:08 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@844 -- # xtrace_disable 00:18:18.049 15:24:08 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@10 -- # set +x 00:18:18.309 [2024-11-19 15:24:08.426386] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:18:18.309 [2024-11-19 15:24:08.427109] [ DPDK EAL parameters: bdevio --no-shconf -c 0x7 -m 0 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100335 ] 00:18:18.309 [2024-11-19 15:24:08.583876] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:18:18.309 [2024-11-19 15:24:08.613108] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:18:18.309 [2024-11-19 15:24:08.613187] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:18:18.309 [2024-11-19 15:24:08.613274] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:18:19.248 15:24:09 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:18:19.248 15:24:09 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@868 -- # return 0 00:18:19.248 15:24:09 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@293 -- # /home/vagrant/spdk_repo/spdk/test/bdev/bdevio/tests.py perform_tests 00:18:19.248 I/O targets: 00:18:19.248 raid5f: 131072 blocks of 512 bytes (64 MiB) 00:18:19.248 00:18:19.248 00:18:19.248 CUnit - A unit testing framework for C - Version 2.1-3 00:18:19.248 http://cunit.sourceforge.net/ 00:18:19.248 00:18:19.248 00:18:19.248 Suite: bdevio tests on: raid5f 00:18:19.248 Test: blockdev write read block ...passed 00:18:19.248 Test: blockdev write zeroes read block ...passed 00:18:19.248 Test: blockdev write zeroes read no split ...passed 00:18:19.248 Test: blockdev write zeroes read split ...passed 00:18:19.248 Test: blockdev write zeroes read split partial ...passed 00:18:19.248 Test: blockdev reset ...passed 00:18:19.248 Test: blockdev write read 8 blocks ...passed 00:18:19.248 Test: blockdev write read size > 128k ...passed 00:18:19.248 Test: blockdev write read invalid size ...passed 00:18:19.248 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:18:19.248 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:18:19.248 Test: blockdev write read max offset ...passed 00:18:19.248 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:18:19.248 Test: blockdev writev readv 8 blocks ...passed 00:18:19.248 Test: blockdev writev readv 30 x 1block ...passed 00:18:19.248 Test: blockdev writev readv block ...passed 00:18:19.248 Test: blockdev writev readv size > 128k ...passed 00:18:19.248 Test: blockdev writev readv size > 128k in two iovs ...passed 00:18:19.248 Test: blockdev comparev and writev ...passed 00:18:19.248 Test: blockdev nvme passthru rw ...passed 00:18:19.248 Test: blockdev nvme passthru vendor specific ...passed 00:18:19.248 Test: blockdev nvme admin passthru ...passed 00:18:19.248 Test: blockdev copy ...passed 00:18:19.248 00:18:19.248 Run Summary: Type Total Ran Passed Failed Inactive 00:18:19.248 suites 1 1 n/a 0 0 00:18:19.248 tests 23 23 23 0 0 00:18:19.248 asserts 130 130 130 0 n/a 00:18:19.248 00:18:19.248 Elapsed time = 0.291 seconds 00:18:19.248 0 00:18:19.248 15:24:09 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@294 -- # killprocess 100335 00:18:19.248 15:24:09 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@954 -- # '[' -z 100335 ']' 00:18:19.248 15:24:09 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@958 -- # kill -0 100335 00:18:19.248 15:24:09 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@959 -- # uname 00:18:19.248 15:24:09 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:18:19.248 15:24:09 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 100335 00:18:19.248 15:24:09 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:18:19.248 15:24:09 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:18:19.248 killing process with pid 100335 00:18:19.248 15:24:09 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@972 -- # echo 'killing process with pid 100335' 00:18:19.248 15:24:09 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@973 -- # kill 100335 00:18:19.248 15:24:09 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@978 -- # wait 100335 00:18:19.509 15:24:09 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@295 -- # trap - SIGINT SIGTERM EXIT 00:18:19.509 00:18:19.509 real 0m1.415s 00:18:19.509 user 0m3.420s 00:18:19.509 sys 0m0.351s 00:18:19.509 ************************************ 00:18:19.509 END TEST bdev_bounds 00:18:19.509 ************************************ 00:18:19.509 15:24:09 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@1130 -- # xtrace_disable 00:18:19.509 15:24:09 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@10 -- # set +x 00:18:19.509 15:24:09 blockdev_raid5f -- bdev/blockdev.sh@761 -- # run_test bdev_nbd nbd_function_test /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json raid5f '' 00:18:19.509 15:24:09 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:18:19.509 15:24:09 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:18:19.509 15:24:09 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:18:19.509 ************************************ 00:18:19.509 START TEST bdev_nbd 00:18:19.509 ************************************ 00:18:19.509 15:24:09 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@1129 -- # nbd_function_test /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json raid5f '' 00:18:19.509 15:24:09 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@299 -- # uname -s 00:18:19.509 15:24:09 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@299 -- # [[ Linux == Linux ]] 00:18:19.509 15:24:09 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@301 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:18:19.509 15:24:09 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@302 -- # local conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json 00:18:19.509 15:24:09 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@303 -- # bdev_all=('raid5f') 00:18:19.509 15:24:09 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@303 -- # local bdev_all 00:18:19.509 15:24:09 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@304 -- # local bdev_num=1 00:18:19.509 15:24:09 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@308 -- # [[ -e /sys/module/nbd ]] 00:18:19.509 15:24:09 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@310 -- # nbd_all=('/dev/nbd0' '/dev/nbd1' '/dev/nbd10' '/dev/nbd11' '/dev/nbd12' '/dev/nbd13' '/dev/nbd14' '/dev/nbd15' '/dev/nbd2' '/dev/nbd3' '/dev/nbd4' '/dev/nbd5' '/dev/nbd6' '/dev/nbd7' '/dev/nbd8' '/dev/nbd9') 00:18:19.509 15:24:09 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@310 -- # local nbd_all 00:18:19.509 15:24:09 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@311 -- # bdev_num=1 00:18:19.509 15:24:09 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@313 -- # nbd_list=('/dev/nbd0') 00:18:19.509 15:24:09 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@313 -- # local nbd_list 00:18:19.509 15:24:09 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@314 -- # bdev_list=('raid5f') 00:18:19.509 15:24:09 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@314 -- # local bdev_list 00:18:19.509 15:24:09 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@317 -- # nbd_pid=100378 00:18:19.509 15:24:09 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@316 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-nbd.sock -i 0 --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json '' 00:18:19.509 15:24:09 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@318 -- # trap 'cleanup; killprocess $nbd_pid' SIGINT SIGTERM EXIT 00:18:19.509 15:24:09 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@319 -- # waitforlisten 100378 /var/tmp/spdk-nbd.sock 00:18:19.509 15:24:09 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@835 -- # '[' -z 100378 ']' 00:18:19.509 15:24:09 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:18:19.509 15:24:09 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@840 -- # local max_retries=100 00:18:19.509 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:18:19.509 15:24:09 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:18:19.509 15:24:09 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@844 -- # xtrace_disable 00:18:19.509 15:24:09 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@10 -- # set +x 00:18:19.770 [2024-11-19 15:24:09.933866] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:18:19.770 [2024-11-19 15:24:09.934106] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:18:19.770 [2024-11-19 15:24:10.090011] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:18:20.041 [2024-11-19 15:24:10.114925] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:18:20.616 15:24:10 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:18:20.616 15:24:10 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@868 -- # return 0 00:18:20.616 15:24:10 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@321 -- # nbd_rpc_start_stop_verify /var/tmp/spdk-nbd.sock raid5f 00:18:20.616 15:24:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@113 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:18:20.616 15:24:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@114 -- # bdev_list=('raid5f') 00:18:20.616 15:24:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@114 -- # local bdev_list 00:18:20.616 15:24:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@116 -- # nbd_start_disks_without_nbd_idx /var/tmp/spdk-nbd.sock raid5f 00:18:20.616 15:24:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@22 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:18:20.616 15:24:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@23 -- # bdev_list=('raid5f') 00:18:20.616 15:24:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@23 -- # local bdev_list 00:18:20.616 15:24:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@24 -- # local i 00:18:20.616 15:24:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@25 -- # local nbd_device 00:18:20.616 15:24:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i = 0 )) 00:18:20.616 15:24:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 1 )) 00:18:20.616 15:24:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk raid5f 00:18:20.876 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@28 -- # nbd_device=/dev/nbd0 00:18:20.876 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@30 -- # basename /dev/nbd0 00:18:20.876 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@30 -- # waitfornbd nbd0 00:18:20.876 15:24:11 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:18:20.876 15:24:11 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@873 -- # local i 00:18:20.876 15:24:11 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:18:20.876 15:24:11 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:18:20.876 15:24:11 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:18:20.876 15:24:11 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@877 -- # break 00:18:20.876 15:24:11 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:18:20.876 15:24:11 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:18:20.876 15:24:11 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:18:20.876 1+0 records in 00:18:20.876 1+0 records out 00:18:20.876 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000308127 s, 13.3 MB/s 00:18:20.876 15:24:11 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:18:20.876 15:24:11 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@890 -- # size=4096 00:18:20.876 15:24:11 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:18:20.876 15:24:11 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:18:20.876 15:24:11 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@893 -- # return 0 00:18:20.876 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i++ )) 00:18:20.876 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 1 )) 00:18:20.876 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@118 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:18:21.137 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@118 -- # nbd_disks_json='[ 00:18:21.137 { 00:18:21.137 "nbd_device": "/dev/nbd0", 00:18:21.137 "bdev_name": "raid5f" 00:18:21.137 } 00:18:21.137 ]' 00:18:21.137 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@119 -- # nbd_disks_name=($(echo "${nbd_disks_json}" | jq -r '.[] | .nbd_device')) 00:18:21.137 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@119 -- # echo '[ 00:18:21.137 { 00:18:21.137 "nbd_device": "/dev/nbd0", 00:18:21.137 "bdev_name": "raid5f" 00:18:21.137 } 00:18:21.137 ]' 00:18:21.137 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@119 -- # jq -r '.[] | .nbd_device' 00:18:21.137 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@120 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock /dev/nbd0 00:18:21.137 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:18:21.137 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:18:21.137 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:18:21.137 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:18:21.137 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:18:21.137 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:18:21.398 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:18:21.398 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:18:21.398 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:18:21.398 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:18:21.398 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:18:21.398 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:18:21.398 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:18:21.398 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:18:21.398 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@122 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:18:21.398 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:18:21.398 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:18:21.398 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:18:21.398 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[]' 00:18:21.398 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:18:21.658 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:18:21.658 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:18:21.658 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo '' 00:18:21.658 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # true 00:18:21.658 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=0 00:18:21.658 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 0 00:18:21.658 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@122 -- # count=0 00:18:21.658 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@123 -- # '[' 0 -ne 0 ']' 00:18:21.658 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@127 -- # return 0 00:18:21.658 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@322 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock raid5f /dev/nbd0 00:18:21.658 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:18:21.658 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@91 -- # bdev_list=('raid5f') 00:18:21.658 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@91 -- # local bdev_list 00:18:21.658 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0') 00:18:21.658 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@92 -- # local nbd_list 00:18:21.658 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock raid5f /dev/nbd0 00:18:21.658 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:18:21.658 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@10 -- # bdev_list=('raid5f') 00:18:21.658 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@10 -- # local bdev_list 00:18:21.658 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:18:21.658 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@11 -- # local nbd_list 00:18:21.659 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@12 -- # local i 00:18:21.659 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:18:21.659 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:18:21.659 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk raid5f /dev/nbd0 00:18:21.659 /dev/nbd0 00:18:21.918 15:24:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:18:21.919 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:18:21.919 15:24:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:18:21.919 15:24:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@873 -- # local i 00:18:21.919 15:24:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:18:21.919 15:24:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:18:21.919 15:24:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:18:21.919 15:24:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@877 -- # break 00:18:21.919 15:24:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:18:21.919 15:24:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:18:21.919 15:24:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:18:21.919 1+0 records in 00:18:21.919 1+0 records out 00:18:21.919 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.0004089 s, 10.0 MB/s 00:18:21.919 15:24:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:18:21.919 15:24:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@890 -- # size=4096 00:18:21.919 15:24:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:18:21.919 15:24:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:18:21.919 15:24:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@893 -- # return 0 00:18:21.919 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:18:21.919 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:18:21.919 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:18:21.919 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:18:21.919 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:18:21.919 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:18:21.919 { 00:18:21.919 "nbd_device": "/dev/nbd0", 00:18:21.919 "bdev_name": "raid5f" 00:18:21.919 } 00:18:21.919 ]' 00:18:21.919 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[ 00:18:21.919 { 00:18:21.919 "nbd_device": "/dev/nbd0", 00:18:21.919 "bdev_name": "raid5f" 00:18:21.919 } 00:18:21.919 ]' 00:18:21.919 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:18:22.179 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name=/dev/nbd0 00:18:22.179 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo /dev/nbd0 00:18:22.179 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:18:22.179 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=1 00:18:22.179 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 1 00:18:22.179 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@95 -- # count=1 00:18:22.179 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@96 -- # '[' 1 -ne 1 ']' 00:18:22.179 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify /dev/nbd0 write 00:18:22.179 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0') 00:18:22.179 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # local nbd_list 00:18:22.179 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@71 -- # local operation=write 00:18:22.179 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest 00:18:22.179 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:18:22.179 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest bs=4096 count=256 00:18:22.179 256+0 records in 00:18:22.179 256+0 records out 00:18:22.179 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0139581 s, 75.1 MB/s 00:18:22.179 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:18:22.179 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:18:22.179 256+0 records in 00:18:22.179 256+0 records out 00:18:22.179 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0271299 s, 38.7 MB/s 00:18:22.179 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify /dev/nbd0 verify 00:18:22.179 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0') 00:18:22.179 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # local nbd_list 00:18:22.179 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@71 -- # local operation=verify 00:18:22.179 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest 00:18:22.179 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:18:22.179 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:18:22.179 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:18:22.179 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest /dev/nbd0 00:18:22.179 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest 00:18:22.179 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock /dev/nbd0 00:18:22.179 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:18:22.179 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:18:22.179 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:18:22.179 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:18:22.179 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:18:22.179 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:18:22.439 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:18:22.439 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:18:22.439 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:18:22.439 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:18:22.439 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:18:22.439 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:18:22.440 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:18:22.440 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:18:22.440 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:18:22.440 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:18:22.440 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:18:22.440 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:18:22.440 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[]' 00:18:22.440 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:18:22.700 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:18:22.700 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo '' 00:18:22.700 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:18:22.700 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # true 00:18:22.700 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=0 00:18:22.700 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 0 00:18:22.700 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@104 -- # count=0 00:18:22.700 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:18:22.700 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@109 -- # return 0 00:18:22.700 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@323 -- # nbd_with_lvol_verify /var/tmp/spdk-nbd.sock /dev/nbd0 00:18:22.700 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@131 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:18:22.700 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@132 -- # local nbd=/dev/nbd0 00:18:22.700 15:24:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@134 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create -b malloc_lvol_verify 16 512 00:18:22.700 malloc_lvol_verify 00:18:22.700 15:24:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@135 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_lvol_create_lvstore malloc_lvol_verify lvs 00:18:22.960 0eb94042-3aeb-442b-b571-af69b29d5f5a 00:18:22.960 15:24:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@136 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_lvol_create lvol 4 -l lvs 00:18:23.220 f3563993-783c-48e1-9d6f-074d43086b2f 00:18:23.220 15:24:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@137 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk lvs/lvol /dev/nbd0 00:18:23.482 /dev/nbd0 00:18:23.482 15:24:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@139 -- # wait_for_nbd_set_capacity /dev/nbd0 00:18:23.482 15:24:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@146 -- # local nbd=nbd0 00:18:23.482 15:24:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@148 -- # [[ -e /sys/block/nbd0/size ]] 00:18:23.482 15:24:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@150 -- # (( 8192 == 0 )) 00:18:23.482 15:24:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@141 -- # mkfs.ext4 /dev/nbd0 00:18:23.482 Discarding device blocks: 0/4096mke2fs 1.47.0 (5-Feb-2023) 00:18:23.482 done 00:18:23.482 Creating filesystem with 4096 1k blocks and 1024 inodes 00:18:23.482 00:18:23.482 Allocating group tables: 0/1 done 00:18:23.482 Writing inode tables: 0/1 done 00:18:23.482 Creating journal (1024 blocks): done 00:18:23.482 Writing superblocks and filesystem accounting information: 0/1 done 00:18:23.482 00:18:23.482 15:24:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@142 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock /dev/nbd0 00:18:23.482 15:24:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:18:23.482 15:24:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:18:23.482 15:24:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:18:23.482 15:24:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:18:23.482 15:24:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:18:23.482 15:24:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:18:23.759 15:24:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:18:23.759 15:24:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:18:23.759 15:24:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:18:23.759 15:24:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:18:23.759 15:24:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:18:23.759 15:24:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:18:23.759 15:24:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:18:23.759 15:24:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:18:23.759 15:24:13 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@325 -- # killprocess 100378 00:18:23.759 15:24:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@954 -- # '[' -z 100378 ']' 00:18:23.759 15:24:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@958 -- # kill -0 100378 00:18:23.759 15:24:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@959 -- # uname 00:18:23.759 15:24:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:18:23.759 15:24:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 100378 00:18:23.759 killing process with pid 100378 00:18:23.759 15:24:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:18:23.759 15:24:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:18:23.759 15:24:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@972 -- # echo 'killing process with pid 100378' 00:18:23.759 15:24:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@973 -- # kill 100378 00:18:23.759 15:24:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@978 -- # wait 100378 00:18:24.037 15:24:14 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@326 -- # trap - SIGINT SIGTERM EXIT 00:18:24.037 00:18:24.037 real 0m4.321s 00:18:24.037 user 0m6.310s 00:18:24.037 sys 0m1.261s 00:18:24.037 15:24:14 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@1130 -- # xtrace_disable 00:18:24.037 15:24:14 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@10 -- # set +x 00:18:24.037 ************************************ 00:18:24.037 END TEST bdev_nbd 00:18:24.037 ************************************ 00:18:24.037 15:24:14 blockdev_raid5f -- bdev/blockdev.sh@762 -- # [[ y == y ]] 00:18:24.037 15:24:14 blockdev_raid5f -- bdev/blockdev.sh@763 -- # '[' raid5f = nvme ']' 00:18:24.037 15:24:14 blockdev_raid5f -- bdev/blockdev.sh@763 -- # '[' raid5f = gpt ']' 00:18:24.037 15:24:14 blockdev_raid5f -- bdev/blockdev.sh@767 -- # run_test bdev_fio fio_test_suite '' 00:18:24.037 15:24:14 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:18:24.037 15:24:14 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:18:24.037 15:24:14 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:18:24.037 ************************************ 00:18:24.037 START TEST bdev_fio 00:18:24.037 ************************************ 00:18:24.037 15:24:14 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1129 -- # fio_test_suite '' 00:18:24.037 15:24:14 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@330 -- # local env_context 00:18:24.037 15:24:14 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@334 -- # pushd /home/vagrant/spdk_repo/spdk/test/bdev 00:18:24.037 /home/vagrant/spdk_repo/spdk/test/bdev /home/vagrant/spdk_repo/spdk 00:18:24.037 15:24:14 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@335 -- # trap 'rm -f ./*.state; popd; exit 1' SIGINT SIGTERM EXIT 00:18:24.037 15:24:14 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@338 -- # echo '' 00:18:24.038 15:24:14 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@338 -- # sed s/--env-context=// 00:18:24.038 15:24:14 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@338 -- # env_context= 00:18:24.038 15:24:14 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@339 -- # fio_config_gen /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio verify AIO '' 00:18:24.038 15:24:14 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1284 -- # local config_file=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:18:24.038 15:24:14 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1285 -- # local workload=verify 00:18:24.038 15:24:14 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1286 -- # local bdev_type=AIO 00:18:24.038 15:24:14 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1287 -- # local env_context= 00:18:24.038 15:24:14 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1288 -- # local fio_dir=/usr/src/fio 00:18:24.038 15:24:14 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1290 -- # '[' -e /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio ']' 00:18:24.038 15:24:14 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1295 -- # '[' -z verify ']' 00:18:24.038 15:24:14 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1299 -- # '[' -n '' ']' 00:18:24.038 15:24:14 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1303 -- # touch /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:18:24.038 15:24:14 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1305 -- # cat 00:18:24.038 15:24:14 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1317 -- # '[' verify == verify ']' 00:18:24.038 15:24:14 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1318 -- # cat 00:18:24.038 15:24:14 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1327 -- # '[' AIO == AIO ']' 00:18:24.038 15:24:14 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1328 -- # /usr/src/fio/fio --version 00:18:24.038 15:24:14 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1328 -- # [[ fio-3.35 == *\f\i\o\-\3* ]] 00:18:24.038 15:24:14 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1329 -- # echo serialize_overlap=1 00:18:24.038 15:24:14 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@340 -- # for b in "${bdevs_name[@]}" 00:18:24.038 15:24:14 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@341 -- # echo '[job_raid5f]' 00:18:24.038 15:24:14 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@342 -- # echo filename=raid5f 00:18:24.038 15:24:14 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@346 -- # local 'fio_params=--ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json' 00:18:24.038 15:24:14 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@348 -- # run_test bdev_fio_rw_verify fio_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:18:24.038 15:24:14 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1105 -- # '[' 11 -le 1 ']' 00:18:24.298 15:24:14 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1111 -- # xtrace_disable 00:18:24.298 15:24:14 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@10 -- # set +x 00:18:24.298 ************************************ 00:18:24.298 START TEST bdev_fio_rw_verify 00:18:24.298 ************************************ 00:18:24.298 15:24:14 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1129 -- # fio_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:18:24.298 15:24:14 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1360 -- # fio_plugin /home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:18:24.298 15:24:14 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1341 -- # local fio_dir=/usr/src/fio 00:18:24.298 15:24:14 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1343 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:18:24.298 15:24:14 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1343 -- # local sanitizers 00:18:24.298 15:24:14 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1344 -- # local plugin=/home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev 00:18:24.298 15:24:14 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # shift 00:18:24.298 15:24:14 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1347 -- # local asan_lib= 00:18:24.298 15:24:14 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1348 -- # for sanitizer in "${sanitizers[@]}" 00:18:24.298 15:24:14 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1349 -- # ldd /home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev 00:18:24.298 15:24:14 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1349 -- # grep libasan 00:18:24.298 15:24:14 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1349 -- # awk '{print $3}' 00:18:24.298 15:24:14 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1349 -- # asan_lib=/usr/lib64/libasan.so.8 00:18:24.298 15:24:14 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1350 -- # [[ -n /usr/lib64/libasan.so.8 ]] 00:18:24.298 15:24:14 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1351 -- # break 00:18:24.298 15:24:14 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1356 -- # LD_PRELOAD='/usr/lib64/libasan.so.8 /home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev' 00:18:24.298 15:24:14 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1356 -- # /usr/src/fio/fio --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:18:24.298 job_raid5f: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:18:24.298 fio-3.35 00:18:24.298 Starting 1 thread 00:18:36.524 00:18:36.524 job_raid5f: (groupid=0, jobs=1): err= 0: pid=100563: Tue Nov 19 15:24:25 2024 00:18:36.524 read: IOPS=12.7k, BW=49.6MiB/s (52.0MB/s)(496MiB/10000msec) 00:18:36.524 slat (nsec): min=16956, max=70389, avg=18390.56, stdev=1641.89 00:18:36.524 clat (usec): min=11, max=308, avg=128.03, stdev=43.37 00:18:36.524 lat (usec): min=29, max=336, avg=146.42, stdev=43.57 00:18:36.524 clat percentiles (usec): 00:18:36.524 | 50.000th=[ 133], 99.000th=[ 208], 99.900th=[ 231], 99.990th=[ 269], 00:18:36.524 | 99.999th=[ 293] 00:18:36.524 write: IOPS=13.3k, BW=52.0MiB/s (54.5MB/s)(514MiB/9878msec); 0 zone resets 00:18:36.524 slat (usec): min=7, max=185, avg=15.85, stdev= 3.42 00:18:36.524 clat (usec): min=56, max=1437, avg=289.40, stdev=40.65 00:18:36.524 lat (usec): min=71, max=1623, avg=305.26, stdev=41.73 00:18:36.524 clat percentiles (usec): 00:18:36.524 | 50.000th=[ 293], 99.000th=[ 363], 99.900th=[ 594], 99.990th=[ 1205], 00:18:36.524 | 99.999th=[ 1336] 00:18:36.524 bw ( KiB/s): min=50088, max=55136, per=98.97%, avg=52683.79, stdev=1874.05, samples=19 00:18:36.524 iops : min=12522, max=13784, avg=13170.95, stdev=468.51, samples=19 00:18:36.524 lat (usec) : 20=0.01%, 50=0.01%, 100=16.73%, 250=39.99%, 500=43.20% 00:18:36.524 lat (usec) : 750=0.04%, 1000=0.02% 00:18:36.524 lat (msec) : 2=0.01% 00:18:36.524 cpu : usr=98.88%, sys=0.41%, ctx=33, majf=0, minf=13449 00:18:36.524 IO depths : 1=7.6%, 2=19.9%, 4=55.1%, 8=17.4%, 16=0.0%, 32=0.0%, >=64=0.0% 00:18:36.524 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:18:36.524 complete : 0=0.0%, 4=90.0%, 8=10.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:18:36.524 issued rwts: total=127005,131461,0,0 short=0,0,0,0 dropped=0,0,0,0 00:18:36.524 latency : target=0, window=0, percentile=100.00%, depth=8 00:18:36.524 00:18:36.525 Run status group 0 (all jobs): 00:18:36.525 READ: bw=49.6MiB/s (52.0MB/s), 49.6MiB/s-49.6MiB/s (52.0MB/s-52.0MB/s), io=496MiB (520MB), run=10000-10000msec 00:18:36.525 WRITE: bw=52.0MiB/s (54.5MB/s), 52.0MiB/s-52.0MiB/s (54.5MB/s-54.5MB/s), io=514MiB (538MB), run=9878-9878msec 00:18:36.525 ----------------------------------------------------- 00:18:36.525 Suppressions used: 00:18:36.525 count bytes template 00:18:36.525 1 7 /usr/src/fio/parse.c 00:18:36.525 513 49248 /usr/src/fio/iolog.c 00:18:36.525 1 8 libtcmalloc_minimal.so 00:18:36.525 1 904 libcrypto.so 00:18:36.525 ----------------------------------------------------- 00:18:36.525 00:18:36.525 00:18:36.525 real 0m11.277s 00:18:36.525 user 0m11.511s 00:18:36.525 sys 0m0.449s 00:18:36.525 15:24:25 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1130 -- # xtrace_disable 00:18:36.525 15:24:25 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@10 -- # set +x 00:18:36.525 ************************************ 00:18:36.525 END TEST bdev_fio_rw_verify 00:18:36.525 ************************************ 00:18:36.525 15:24:25 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@349 -- # rm -f 00:18:36.525 15:24:25 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@350 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:18:36.525 15:24:25 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@353 -- # fio_config_gen /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio trim '' '' 00:18:36.525 15:24:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1284 -- # local config_file=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:18:36.525 15:24:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1285 -- # local workload=trim 00:18:36.525 15:24:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1286 -- # local bdev_type= 00:18:36.525 15:24:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1287 -- # local env_context= 00:18:36.525 15:24:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1288 -- # local fio_dir=/usr/src/fio 00:18:36.525 15:24:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1290 -- # '[' -e /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio ']' 00:18:36.525 15:24:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1295 -- # '[' -z trim ']' 00:18:36.525 15:24:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1299 -- # '[' -n '' ']' 00:18:36.525 15:24:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1303 -- # touch /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:18:36.525 15:24:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1305 -- # cat 00:18:36.525 15:24:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1317 -- # '[' trim == verify ']' 00:18:36.525 15:24:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1332 -- # '[' trim == trim ']' 00:18:36.525 15:24:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1333 -- # echo rw=trimwrite 00:18:36.525 15:24:25 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@354 -- # printf '%s\n' '{' ' "name": "raid5f",' ' "aliases": [' ' "9420d3b3-3065-48e8-a5e5-7b2bc31c6b72"' ' ],' ' "product_name": "Raid Volume",' ' "block_size": 512,' ' "num_blocks": 131072,' ' "uuid": "9420d3b3-3065-48e8-a5e5-7b2bc31c6b72",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": false,' ' "flush": false,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "raid": {' ' "uuid": "9420d3b3-3065-48e8-a5e5-7b2bc31c6b72",' ' "strip_size_kb": 2,' ' "state": "online",' ' "raid_level": "raid5f",' ' "superblock": false,' ' "num_base_bdevs": 3,' ' "num_base_bdevs_discovered": 3,' ' "num_base_bdevs_operational": 3,' ' "base_bdevs_list": [' ' {' ' "name": "Malloc0",' ' "uuid": "f14beae4-660e-4035-9bac-d91415d8c490",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc1",' ' "uuid": "87a81309-869b-4719-b95c-29beb05627f6",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc2",' ' "uuid": "a835ec76-681d-46c6-acec-65e504f5880d",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' }' ' ]' ' }' ' }' '}' 00:18:36.525 15:24:25 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@354 -- # jq -r 'select(.supported_io_types.unmap == true) | .name' 00:18:36.525 15:24:25 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@354 -- # [[ -n '' ]] 00:18:36.525 15:24:25 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@360 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:18:36.525 15:24:25 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@361 -- # popd 00:18:36.525 /home/vagrant/spdk_repo/spdk 00:18:36.525 15:24:25 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@362 -- # trap - SIGINT SIGTERM EXIT 00:18:36.525 ************************************ 00:18:36.525 END TEST bdev_fio 00:18:36.525 ************************************ 00:18:36.525 15:24:25 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@363 -- # return 0 00:18:36.525 00:18:36.525 real 0m11.584s 00:18:36.525 user 0m11.641s 00:18:36.525 sys 0m0.592s 00:18:36.525 15:24:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1130 -- # xtrace_disable 00:18:36.525 15:24:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@10 -- # set +x 00:18:36.525 15:24:25 blockdev_raid5f -- bdev/blockdev.sh@774 -- # trap cleanup SIGINT SIGTERM EXIT 00:18:36.525 15:24:25 blockdev_raid5f -- bdev/blockdev.sh@776 -- # run_test bdev_verify /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w verify -t 5 -C -m 0x3 '' 00:18:36.525 15:24:25 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 16 -le 1 ']' 00:18:36.525 15:24:25 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:18:36.525 15:24:25 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:18:36.525 ************************************ 00:18:36.525 START TEST bdev_verify 00:18:36.525 ************************************ 00:18:36.525 15:24:25 blockdev_raid5f.bdev_verify -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w verify -t 5 -C -m 0x3 '' 00:18:36.525 [2024-11-19 15:24:25.957417] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:18:36.525 [2024-11-19 15:24:25.957520] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100721 ] 00:18:36.525 [2024-11-19 15:24:26.113786] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:18:36.525 [2024-11-19 15:24:26.160144] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:18:36.525 [2024-11-19 15:24:26.160246] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:18:36.525 Running I/O for 5 seconds... 00:18:38.404 11062.00 IOPS, 43.21 MiB/s [2024-11-19T15:24:29.680Z] 11072.00 IOPS, 43.25 MiB/s [2024-11-19T15:24:30.619Z] 11107.67 IOPS, 43.39 MiB/s [2024-11-19T15:24:31.559Z] 11129.50 IOPS, 43.47 MiB/s [2024-11-19T15:24:31.559Z] 11139.40 IOPS, 43.51 MiB/s 00:18:41.220 Latency(us) 00:18:41.220 [2024-11-19T15:24:31.559Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:18:41.220 Job: raid5f (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:18:41.220 Verification LBA range: start 0x0 length 0x2000 00:18:41.220 raid5f : 5.02 6660.26 26.02 0.00 0.00 28875.09 115.37 20719.68 00:18:41.220 Job: raid5f (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:18:41.220 Verification LBA range: start 0x2000 length 0x2000 00:18:41.220 raid5f : 5.03 4499.15 17.57 0.00 0.00 42750.66 248.62 30678.86 00:18:41.220 [2024-11-19T15:24:31.559Z] =================================================================================================================== 00:18:41.220 [2024-11-19T15:24:31.559Z] Total : 11159.41 43.59 0.00 0.00 34471.50 115.37 30678.86 00:18:41.480 00:18:41.480 real 0m5.924s 00:18:41.480 user 0m10.977s 00:18:41.480 sys 0m0.307s 00:18:41.480 15:24:31 blockdev_raid5f.bdev_verify -- common/autotest_common.sh@1130 -- # xtrace_disable 00:18:41.480 15:24:31 blockdev_raid5f.bdev_verify -- common/autotest_common.sh@10 -- # set +x 00:18:41.480 ************************************ 00:18:41.480 END TEST bdev_verify 00:18:41.480 ************************************ 00:18:41.741 15:24:31 blockdev_raid5f -- bdev/blockdev.sh@777 -- # run_test bdev_verify_big_io /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 65536 -w verify -t 5 -C -m 0x3 '' 00:18:41.741 15:24:31 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 16 -le 1 ']' 00:18:41.741 15:24:31 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:18:41.741 15:24:31 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:18:41.741 ************************************ 00:18:41.741 START TEST bdev_verify_big_io 00:18:41.741 ************************************ 00:18:41.741 15:24:31 blockdev_raid5f.bdev_verify_big_io -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 65536 -w verify -t 5 -C -m 0x3 '' 00:18:41.741 [2024-11-19 15:24:31.965373] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:18:41.741 [2024-11-19 15:24:31.965501] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100798 ] 00:18:42.001 [2024-11-19 15:24:32.122954] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:18:42.001 [2024-11-19 15:24:32.175186] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:18:42.001 [2024-11-19 15:24:32.175283] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:18:42.261 Running I/O for 5 seconds... 00:18:44.148 633.00 IOPS, 39.56 MiB/s [2024-11-19T15:24:35.866Z] 761.00 IOPS, 47.56 MiB/s [2024-11-19T15:24:36.805Z] 803.00 IOPS, 50.19 MiB/s [2024-11-19T15:24:37.745Z] 793.25 IOPS, 49.58 MiB/s [2024-11-19T15:24:37.745Z] 812.00 IOPS, 50.75 MiB/s 00:18:47.406 Latency(us) 00:18:47.406 [2024-11-19T15:24:37.746Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:18:47.407 Job: raid5f (Core Mask 0x1, workload: verify, depth: 128, IO size: 65536) 00:18:47.407 Verification LBA range: start 0x0 length 0x200 00:18:47.407 raid5f : 5.18 465.63 29.10 0.00 0.00 6881641.82 202.12 298546.53 00:18:47.407 Job: raid5f (Core Mask 0x2, workload: verify, depth: 128, IO size: 65536) 00:18:47.407 Verification LBA range: start 0x200 length 0x200 00:18:47.407 raid5f : 5.27 361.23 22.58 0.00 0.00 8751269.67 214.64 377304.20 00:18:47.407 [2024-11-19T15:24:37.746Z] =================================================================================================================== 00:18:47.407 [2024-11-19T15:24:37.746Z] Total : 826.87 51.68 0.00 0.00 7706477.64 202.12 377304.20 00:18:47.976 00:18:47.976 real 0m6.178s 00:18:47.976 user 0m11.449s 00:18:47.976 sys 0m0.323s 00:18:47.976 15:24:38 blockdev_raid5f.bdev_verify_big_io -- common/autotest_common.sh@1130 -- # xtrace_disable 00:18:47.976 15:24:38 blockdev_raid5f.bdev_verify_big_io -- common/autotest_common.sh@10 -- # set +x 00:18:47.976 ************************************ 00:18:47.976 END TEST bdev_verify_big_io 00:18:47.976 ************************************ 00:18:47.976 15:24:38 blockdev_raid5f -- bdev/blockdev.sh@778 -- # run_test bdev_write_zeroes /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:18:47.976 15:24:38 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 13 -le 1 ']' 00:18:47.976 15:24:38 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:18:47.976 15:24:38 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:18:47.977 ************************************ 00:18:47.977 START TEST bdev_write_zeroes 00:18:47.977 ************************************ 00:18:47.977 15:24:38 blockdev_raid5f.bdev_write_zeroes -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:18:47.977 [2024-11-19 15:24:38.213206] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:18:47.977 [2024-11-19 15:24:38.213336] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100887 ] 00:18:48.236 [2024-11-19 15:24:38.369850] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:18:48.236 [2024-11-19 15:24:38.409331] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:18:48.496 Running I/O for 1 seconds... 00:18:49.436 29655.00 IOPS, 115.84 MiB/s 00:18:49.436 Latency(us) 00:18:49.436 [2024-11-19T15:24:39.775Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:18:49.436 Job: raid5f (Core Mask 0x1, workload: write_zeroes, depth: 128, IO size: 4096) 00:18:49.436 raid5f : 1.01 29640.92 115.78 0.00 0.00 4305.09 1366.53 5809.52 00:18:49.436 [2024-11-19T15:24:39.775Z] =================================================================================================================== 00:18:49.436 [2024-11-19T15:24:39.775Z] Total : 29640.92 115.78 0.00 0.00 4305.09 1366.53 5809.52 00:18:49.696 00:18:49.696 real 0m1.894s 00:18:49.696 user 0m1.518s 00:18:49.696 sys 0m0.264s 00:18:49.696 15:24:40 blockdev_raid5f.bdev_write_zeroes -- common/autotest_common.sh@1130 -- # xtrace_disable 00:18:49.696 15:24:40 blockdev_raid5f.bdev_write_zeroes -- common/autotest_common.sh@10 -- # set +x 00:18:49.696 ************************************ 00:18:49.696 END TEST bdev_write_zeroes 00:18:49.696 ************************************ 00:18:49.956 15:24:40 blockdev_raid5f -- bdev/blockdev.sh@781 -- # run_test bdev_json_nonenclosed /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonenclosed.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:18:49.956 15:24:40 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 13 -le 1 ']' 00:18:49.956 15:24:40 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:18:49.956 15:24:40 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:18:49.956 ************************************ 00:18:49.956 START TEST bdev_json_nonenclosed 00:18:49.956 ************************************ 00:18:49.956 15:24:40 blockdev_raid5f.bdev_json_nonenclosed -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonenclosed.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:18:49.956 [2024-11-19 15:24:40.188934] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:18:49.956 [2024-11-19 15:24:40.189070] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100923 ] 00:18:50.216 [2024-11-19 15:24:40.345771] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:18:50.216 [2024-11-19 15:24:40.391545] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:18:50.216 [2024-11-19 15:24:40.391660] json_config.c: 608:json_config_prepare_ctx: *ERROR*: Invalid JSON configuration: not enclosed in {}. 00:18:50.216 [2024-11-19 15:24:40.391682] rpc.c: 190:spdk_rpc_server_finish: *ERROR*: No server listening on provided address: 00:18:50.216 [2024-11-19 15:24:40.391696] app.c:1064:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:18:50.216 00:18:50.216 real 0m0.397s 00:18:50.216 user 0m0.165s 00:18:50.216 sys 0m0.128s 00:18:50.216 15:24:40 blockdev_raid5f.bdev_json_nonenclosed -- common/autotest_common.sh@1130 -- # xtrace_disable 00:18:50.216 15:24:40 blockdev_raid5f.bdev_json_nonenclosed -- common/autotest_common.sh@10 -- # set +x 00:18:50.216 ************************************ 00:18:50.216 END TEST bdev_json_nonenclosed 00:18:50.216 ************************************ 00:18:50.475 15:24:40 blockdev_raid5f -- bdev/blockdev.sh@784 -- # run_test bdev_json_nonarray /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonarray.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:18:50.475 15:24:40 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 13 -le 1 ']' 00:18:50.475 15:24:40 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:18:50.475 15:24:40 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:18:50.475 ************************************ 00:18:50.475 START TEST bdev_json_nonarray 00:18:50.475 ************************************ 00:18:50.475 15:24:40 blockdev_raid5f.bdev_json_nonarray -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonarray.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:18:50.475 [2024-11-19 15:24:40.654278] Starting SPDK v25.01-pre git sha1 dcc2ca8f3 / DPDK 22.11.4 initialization... 00:18:50.475 [2024-11-19 15:24:40.654394] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100953 ] 00:18:50.475 [2024-11-19 15:24:40.809791] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:18:50.736 [2024-11-19 15:24:40.850195] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:18:50.736 [2024-11-19 15:24:40.850328] json_config.c: 614:json_config_prepare_ctx: *ERROR*: Invalid JSON configuration: 'subsystems' should be an array. 00:18:50.736 [2024-11-19 15:24:40.850354] rpc.c: 190:spdk_rpc_server_finish: *ERROR*: No server listening on provided address: 00:18:50.736 [2024-11-19 15:24:40.850366] app.c:1064:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:18:50.736 00:18:50.736 real 0m0.383s 00:18:50.736 user 0m0.166s 00:18:50.736 sys 0m0.113s 00:18:50.736 15:24:40 blockdev_raid5f.bdev_json_nonarray -- common/autotest_common.sh@1130 -- # xtrace_disable 00:18:50.736 15:24:40 blockdev_raid5f.bdev_json_nonarray -- common/autotest_common.sh@10 -- # set +x 00:18:50.736 ************************************ 00:18:50.736 END TEST bdev_json_nonarray 00:18:50.736 ************************************ 00:18:50.736 15:24:41 blockdev_raid5f -- bdev/blockdev.sh@786 -- # [[ raid5f == bdev ]] 00:18:50.736 15:24:41 blockdev_raid5f -- bdev/blockdev.sh@793 -- # [[ raid5f == gpt ]] 00:18:50.736 15:24:41 blockdev_raid5f -- bdev/blockdev.sh@797 -- # [[ raid5f == crypto_sw ]] 00:18:50.736 15:24:41 blockdev_raid5f -- bdev/blockdev.sh@809 -- # trap - SIGINT SIGTERM EXIT 00:18:50.736 15:24:41 blockdev_raid5f -- bdev/blockdev.sh@810 -- # cleanup 00:18:50.736 15:24:41 blockdev_raid5f -- bdev/blockdev.sh@23 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/aiofile 00:18:50.736 15:24:41 blockdev_raid5f -- bdev/blockdev.sh@24 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json 00:18:50.736 15:24:41 blockdev_raid5f -- bdev/blockdev.sh@26 -- # [[ raid5f == rbd ]] 00:18:50.736 15:24:41 blockdev_raid5f -- bdev/blockdev.sh@30 -- # [[ raid5f == daos ]] 00:18:50.736 15:24:41 blockdev_raid5f -- bdev/blockdev.sh@34 -- # [[ raid5f = \g\p\t ]] 00:18:50.736 15:24:41 blockdev_raid5f -- bdev/blockdev.sh@40 -- # [[ raid5f == xnvme ]] 00:18:50.736 00:18:50.736 real 0m35.280s 00:18:50.736 user 0m47.989s 00:18:50.736 sys 0m4.593s 00:18:50.736 15:24:41 blockdev_raid5f -- common/autotest_common.sh@1130 -- # xtrace_disable 00:18:50.736 15:24:41 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:18:50.736 ************************************ 00:18:50.736 END TEST blockdev_raid5f 00:18:50.736 ************************************ 00:18:50.997 15:24:41 -- spdk/autotest.sh@194 -- # uname -s 00:18:50.997 15:24:41 -- spdk/autotest.sh@194 -- # [[ Linux == Linux ]] 00:18:50.997 15:24:41 -- spdk/autotest.sh@195 -- # [[ 0 -eq 1 ]] 00:18:50.997 15:24:41 -- spdk/autotest.sh@195 -- # [[ 0 -eq 1 ]] 00:18:50.997 15:24:41 -- spdk/autotest.sh@207 -- # '[' 0 -eq 1 ']' 00:18:50.997 15:24:41 -- spdk/autotest.sh@256 -- # '[' 0 -eq 1 ']' 00:18:50.997 15:24:41 -- spdk/autotest.sh@260 -- # timing_exit lib 00:18:50.997 15:24:41 -- common/autotest_common.sh@732 -- # xtrace_disable 00:18:50.997 15:24:41 -- common/autotest_common.sh@10 -- # set +x 00:18:50.997 15:24:41 -- spdk/autotest.sh@262 -- # '[' 0 -eq 1 ']' 00:18:50.997 15:24:41 -- spdk/autotest.sh@267 -- # '[' 0 -eq 1 ']' 00:18:50.997 15:24:41 -- spdk/autotest.sh@276 -- # '[' 0 -eq 1 ']' 00:18:50.997 15:24:41 -- spdk/autotest.sh@311 -- # '[' 0 -eq 1 ']' 00:18:50.997 15:24:41 -- spdk/autotest.sh@315 -- # '[' 0 -eq 1 ']' 00:18:50.997 15:24:41 -- spdk/autotest.sh@319 -- # '[' 0 -eq 1 ']' 00:18:50.997 15:24:41 -- spdk/autotest.sh@324 -- # '[' 0 -eq 1 ']' 00:18:50.997 15:24:41 -- spdk/autotest.sh@333 -- # '[' 0 -eq 1 ']' 00:18:50.997 15:24:41 -- spdk/autotest.sh@338 -- # '[' 0 -eq 1 ']' 00:18:50.997 15:24:41 -- spdk/autotest.sh@342 -- # '[' 0 -eq 1 ']' 00:18:50.997 15:24:41 -- spdk/autotest.sh@346 -- # '[' 0 -eq 1 ']' 00:18:50.997 15:24:41 -- spdk/autotest.sh@350 -- # '[' 0 -eq 1 ']' 00:18:50.997 15:24:41 -- spdk/autotest.sh@355 -- # '[' 0 -eq 1 ']' 00:18:50.997 15:24:41 -- spdk/autotest.sh@359 -- # '[' 0 -eq 1 ']' 00:18:50.997 15:24:41 -- spdk/autotest.sh@366 -- # [[ 0 -eq 1 ]] 00:18:50.997 15:24:41 -- spdk/autotest.sh@370 -- # [[ 0 -eq 1 ]] 00:18:50.997 15:24:41 -- spdk/autotest.sh@374 -- # [[ 0 -eq 1 ]] 00:18:50.997 15:24:41 -- spdk/autotest.sh@378 -- # [[ '' -eq 1 ]] 00:18:50.997 15:24:41 -- spdk/autotest.sh@385 -- # trap - SIGINT SIGTERM EXIT 00:18:50.997 15:24:41 -- spdk/autotest.sh@387 -- # timing_enter post_cleanup 00:18:50.997 15:24:41 -- common/autotest_common.sh@726 -- # xtrace_disable 00:18:50.997 15:24:41 -- common/autotest_common.sh@10 -- # set +x 00:18:50.997 15:24:41 -- spdk/autotest.sh@388 -- # autotest_cleanup 00:18:50.997 15:24:41 -- common/autotest_common.sh@1396 -- # local autotest_es=0 00:18:50.997 15:24:41 -- common/autotest_common.sh@1397 -- # xtrace_disable 00:18:50.997 15:24:41 -- common/autotest_common.sh@10 -- # set +x 00:18:53.540 INFO: APP EXITING 00:18:53.540 INFO: killing all VMs 00:18:53.540 INFO: killing vhost app 00:18:53.540 INFO: EXIT DONE 00:18:53.800 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:18:53.800 Waiting for block devices as requested 00:18:53.800 0000:00:11.0 (1b36 0010): uio_pci_generic -> nvme 00:18:54.065 0000:00:10.0 (1b36 0010): uio_pci_generic -> nvme 00:18:55.051 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:18:55.051 Cleaning 00:18:55.051 Removing: /var/run/dpdk/spdk0/config 00:18:55.051 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-0 00:18:55.051 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-1 00:18:55.051 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-2 00:18:55.051 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-3 00:18:55.051 Removing: /var/run/dpdk/spdk0/fbarray_memzone 00:18:55.051 Removing: /var/run/dpdk/spdk0/hugepage_info 00:18:55.051 Removing: /dev/shm/spdk_tgt_trace.pid69008 00:18:55.051 Removing: /var/run/dpdk/spdk0 00:18:55.051 Removing: /var/run/dpdk/spdk_pid100012 00:18:55.051 Removing: /var/run/dpdk/spdk_pid100264 00:18:55.051 Removing: /var/run/dpdk/spdk_pid100304 00:18:55.051 Removing: /var/run/dpdk/spdk_pid100335 00:18:55.051 Removing: /var/run/dpdk/spdk_pid100554 00:18:55.051 Removing: /var/run/dpdk/spdk_pid100721 00:18:55.051 Removing: /var/run/dpdk/spdk_pid100798 00:18:55.051 Removing: /var/run/dpdk/spdk_pid100887 00:18:55.051 Removing: /var/run/dpdk/spdk_pid100923 00:18:55.051 Removing: /var/run/dpdk/spdk_pid100953 00:18:55.051 Removing: /var/run/dpdk/spdk_pid68834 00:18:55.051 Removing: /var/run/dpdk/spdk_pid69008 00:18:55.051 Removing: /var/run/dpdk/spdk_pid69215 00:18:55.051 Removing: /var/run/dpdk/spdk_pid69303 00:18:55.051 Removing: /var/run/dpdk/spdk_pid69337 00:18:55.051 Removing: /var/run/dpdk/spdk_pid69448 00:18:55.051 Removing: /var/run/dpdk/spdk_pid69461 00:18:55.051 Removing: /var/run/dpdk/spdk_pid69649 00:18:55.051 Removing: /var/run/dpdk/spdk_pid69735 00:18:55.051 Removing: /var/run/dpdk/spdk_pid69824 00:18:55.051 Removing: /var/run/dpdk/spdk_pid69925 00:18:55.051 Removing: /var/run/dpdk/spdk_pid70006 00:18:55.051 Removing: /var/run/dpdk/spdk_pid70045 00:18:55.051 Removing: /var/run/dpdk/spdk_pid70076 00:18:55.051 Removing: /var/run/dpdk/spdk_pid70152 00:18:55.051 Removing: /var/run/dpdk/spdk_pid70265 00:18:55.051 Removing: /var/run/dpdk/spdk_pid70696 00:18:55.051 Removing: /var/run/dpdk/spdk_pid70745 00:18:55.051 Removing: /var/run/dpdk/spdk_pid70797 00:18:55.051 Removing: /var/run/dpdk/spdk_pid70813 00:18:55.051 Removing: /var/run/dpdk/spdk_pid70882 00:18:55.051 Removing: /var/run/dpdk/spdk_pid70898 00:18:55.051 Removing: /var/run/dpdk/spdk_pid70969 00:18:55.051 Removing: /var/run/dpdk/spdk_pid70985 00:18:55.051 Removing: /var/run/dpdk/spdk_pid71038 00:18:55.051 Removing: /var/run/dpdk/spdk_pid71056 00:18:55.051 Removing: /var/run/dpdk/spdk_pid71098 00:18:55.311 Removing: /var/run/dpdk/spdk_pid71116 00:18:55.311 Removing: /var/run/dpdk/spdk_pid71254 00:18:55.311 Removing: /var/run/dpdk/spdk_pid71291 00:18:55.311 Removing: /var/run/dpdk/spdk_pid71374 00:18:55.311 Removing: /var/run/dpdk/spdk_pid72564 00:18:55.311 Removing: /var/run/dpdk/spdk_pid72770 00:18:55.311 Removing: /var/run/dpdk/spdk_pid72899 00:18:55.311 Removing: /var/run/dpdk/spdk_pid73504 00:18:55.311 Removing: /var/run/dpdk/spdk_pid73704 00:18:55.311 Removing: /var/run/dpdk/spdk_pid73833 00:18:55.311 Removing: /var/run/dpdk/spdk_pid74449 00:18:55.311 Removing: /var/run/dpdk/spdk_pid74763 00:18:55.311 Removing: /var/run/dpdk/spdk_pid74897 00:18:55.311 Removing: /var/run/dpdk/spdk_pid76233 00:18:55.311 Removing: /var/run/dpdk/spdk_pid76475 00:18:55.311 Removing: /var/run/dpdk/spdk_pid76609 00:18:55.311 Removing: /var/run/dpdk/spdk_pid77950 00:18:55.311 Removing: /var/run/dpdk/spdk_pid78192 00:18:55.311 Removing: /var/run/dpdk/spdk_pid78321 00:18:55.311 Removing: /var/run/dpdk/spdk_pid79662 00:18:55.311 Removing: /var/run/dpdk/spdk_pid80097 00:18:55.311 Removing: /var/run/dpdk/spdk_pid80226 00:18:55.311 Removing: /var/run/dpdk/spdk_pid81670 00:18:55.311 Removing: /var/run/dpdk/spdk_pid81918 00:18:55.311 Removing: /var/run/dpdk/spdk_pid82048 00:18:55.311 Removing: /var/run/dpdk/spdk_pid83489 00:18:55.311 Removing: /var/run/dpdk/spdk_pid83743 00:18:55.311 Removing: /var/run/dpdk/spdk_pid83873 00:18:55.312 Removing: /var/run/dpdk/spdk_pid85314 00:18:55.312 Removing: /var/run/dpdk/spdk_pid85785 00:18:55.312 Removing: /var/run/dpdk/spdk_pid85914 00:18:55.312 Removing: /var/run/dpdk/spdk_pid86047 00:18:55.312 Removing: /var/run/dpdk/spdk_pid86460 00:18:55.312 Removing: /var/run/dpdk/spdk_pid87176 00:18:55.312 Removing: /var/run/dpdk/spdk_pid87534 00:18:55.312 Removing: /var/run/dpdk/spdk_pid88202 00:18:55.312 Removing: /var/run/dpdk/spdk_pid88627 00:18:55.312 Removing: /var/run/dpdk/spdk_pid89368 00:18:55.312 Removing: /var/run/dpdk/spdk_pid89761 00:18:55.312 Removing: /var/run/dpdk/spdk_pid91671 00:18:55.312 Removing: /var/run/dpdk/spdk_pid92099 00:18:55.312 Removing: /var/run/dpdk/spdk_pid92522 00:18:55.312 Removing: /var/run/dpdk/spdk_pid94555 00:18:55.312 Removing: /var/run/dpdk/spdk_pid95024 00:18:55.312 Removing: /var/run/dpdk/spdk_pid95530 00:18:55.312 Removing: /var/run/dpdk/spdk_pid96564 00:18:55.312 Removing: /var/run/dpdk/spdk_pid96874 00:18:55.312 Removing: /var/run/dpdk/spdk_pid97792 00:18:55.312 Removing: /var/run/dpdk/spdk_pid98104 00:18:55.312 Removing: /var/run/dpdk/spdk_pid99031 00:18:55.312 Removing: /var/run/dpdk/spdk_pid99348 00:18:55.312 Clean 00:18:55.572 15:24:45 -- common/autotest_common.sh@1453 -- # return 0 00:18:55.572 15:24:45 -- spdk/autotest.sh@389 -- # timing_exit post_cleanup 00:18:55.572 15:24:45 -- common/autotest_common.sh@732 -- # xtrace_disable 00:18:55.572 15:24:45 -- common/autotest_common.sh@10 -- # set +x 00:18:55.572 15:24:45 -- spdk/autotest.sh@391 -- # timing_exit autotest 00:18:55.572 15:24:45 -- common/autotest_common.sh@732 -- # xtrace_disable 00:18:55.572 15:24:45 -- common/autotest_common.sh@10 -- # set +x 00:18:55.572 15:24:45 -- spdk/autotest.sh@392 -- # chmod a+r /home/vagrant/spdk_repo/spdk/../output/timing.txt 00:18:55.572 15:24:45 -- spdk/autotest.sh@394 -- # [[ -f /home/vagrant/spdk_repo/spdk/../output/udev.log ]] 00:18:55.572 15:24:45 -- spdk/autotest.sh@394 -- # rm -f /home/vagrant/spdk_repo/spdk/../output/udev.log 00:18:55.572 15:24:45 -- spdk/autotest.sh@396 -- # [[ y == y ]] 00:18:55.572 15:24:45 -- spdk/autotest.sh@398 -- # hostname 00:18:55.572 15:24:45 -- spdk/autotest.sh@398 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -c --no-external -d /home/vagrant/spdk_repo/spdk -t fedora39-cloud-1721788873-2326 -o /home/vagrant/spdk_repo/spdk/../output/cov_test.info 00:18:55.832 geninfo: WARNING: invalid characters removed from testname! 00:19:22.402 15:25:08 -- spdk/autotest.sh@399 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -a /home/vagrant/spdk_repo/spdk/../output/cov_base.info -a /home/vagrant/spdk_repo/spdk/../output/cov_test.info -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:19:22.402 15:25:11 -- spdk/autotest.sh@400 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/dpdk/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:19:23.343 15:25:13 -- spdk/autotest.sh@404 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info --ignore-errors unused,unused '/usr/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:19:25.885 15:25:15 -- spdk/autotest.sh@405 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/examples/vmd/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:19:27.795 15:25:18 -- spdk/autotest.sh@406 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/app/spdk_lspci/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:19:29.704 15:25:20 -- spdk/autotest.sh@407 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/app/spdk_top/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:19:32.245 15:25:22 -- spdk/autotest.sh@408 -- # rm -f cov_base.info cov_test.info OLD_STDOUT OLD_STDERR 00:19:32.245 15:25:22 -- spdk/autorun.sh@1 -- $ timing_finish 00:19:32.245 15:25:22 -- common/autotest_common.sh@738 -- $ [[ -e /home/vagrant/spdk_repo/spdk/../output/timing.txt ]] 00:19:32.245 15:25:22 -- common/autotest_common.sh@740 -- $ flamegraph=/usr/local/FlameGraph/flamegraph.pl 00:19:32.245 15:25:22 -- common/autotest_common.sh@741 -- $ [[ -x /usr/local/FlameGraph/flamegraph.pl ]] 00:19:32.245 15:25:22 -- common/autotest_common.sh@744 -- $ /usr/local/FlameGraph/flamegraph.pl --title 'Build Timing' --nametype Step: --countname seconds /home/vagrant/spdk_repo/spdk/../output/timing.txt 00:19:32.245 + [[ -n 6164 ]] 00:19:32.245 + sudo kill 6164 00:19:32.255 [Pipeline] } 00:19:32.271 [Pipeline] // timeout 00:19:32.276 [Pipeline] } 00:19:32.291 [Pipeline] // stage 00:19:32.296 [Pipeline] } 00:19:32.311 [Pipeline] // catchError 00:19:32.321 [Pipeline] stage 00:19:32.323 [Pipeline] { (Stop VM) 00:19:32.336 [Pipeline] sh 00:19:32.620 + vagrant halt 00:19:35.161 ==> default: Halting domain... 00:19:43.307 [Pipeline] sh 00:19:43.638 + vagrant destroy -f 00:19:46.181 ==> default: Removing domain... 00:19:46.194 [Pipeline] sh 00:19:46.479 + mv output /var/jenkins/workspace/raid-vg-autotest/output 00:19:46.490 [Pipeline] } 00:19:46.506 [Pipeline] // stage 00:19:46.511 [Pipeline] } 00:19:46.525 [Pipeline] // dir 00:19:46.531 [Pipeline] } 00:19:46.547 [Pipeline] // wrap 00:19:46.553 [Pipeline] } 00:19:46.567 [Pipeline] // catchError 00:19:46.576 [Pipeline] stage 00:19:46.579 [Pipeline] { (Epilogue) 00:19:46.592 [Pipeline] sh 00:19:46.878 + jbp/jenkins/jjb-config/jobs/scripts/compress_artifacts.sh 00:19:51.091 [Pipeline] catchError 00:19:51.093 [Pipeline] { 00:19:51.105 [Pipeline] sh 00:19:51.391 + jbp/jenkins/jjb-config/jobs/scripts/check_artifacts_size.sh 00:19:51.391 Artifacts sizes are good 00:19:51.401 [Pipeline] } 00:19:51.416 [Pipeline] // catchError 00:19:51.428 [Pipeline] archiveArtifacts 00:19:51.437 Archiving artifacts 00:19:51.543 [Pipeline] cleanWs 00:19:51.556 [WS-CLEANUP] Deleting project workspace... 00:19:51.556 [WS-CLEANUP] Deferred wipeout is used... 00:19:51.563 [WS-CLEANUP] done 00:19:51.565 [Pipeline] } 00:19:51.581 [Pipeline] // stage 00:19:51.586 [Pipeline] } 00:19:51.601 [Pipeline] // node 00:19:51.607 [Pipeline] End of Pipeline 00:19:51.657 Finished: SUCCESS